throbber
(12) Un1ted States Patent
`(10) Patent No.:
`US 7,631,811 B1
`
`Brown
`(45) Date of Patent:
`Dec. 15, 2009
`
`USOO7631811B1
`
`(54) OPTICAL HEADSET USER INTERFACE
`
`(75)
`
`Inventor: William Owen Brown, Santa Cruz, CA
`(US)
`
`(73) Assignee: Plantronics, Inc., Santa Cruz, CA (US)
`
`( * ) Notice:
`
`Subject to any disclaimer, the term of this
`patent is extended or adjusted under 35
`U.S.C. 154(b) by 103 days.
`
`(21) Appl. No.: 11/906,803
`
`(22)
`
`Filed:
`
`Oct. 4, 2007
`
`(51)
`
`Int. Cl.
`(200601)
`G06K 7/14
`(52) US. Cl.
`..................... 235/454; 379/428.02; 455/73
`(58) Field of Classification Search ................. 235/454;
`.
`.
`.379/428-02
`See apphcatron file for complete search hlstory.
`References Cited
`
`(56)
`
`U.S. PATENT DOCUMENTS
`6,980,673 B2 * 12/2005 Funahashi ................... 382/124
`2001/0017934 A1*
`8/2001 Paloniemi et a1.
`........... 382/107
`
`..................... 381/74
`2007/0274530 A1* 11/2007 Buil et al.
`2008/0130910 A1
`6/2008 Jobling et al.
`2008/0284734 A1* 11/2008 Visser
`........................ 345/166
`OTHER PUBLICATIONS
`
`Gregory, Peter; Doria, Tom; Stegh, Chris; Su, Jim; SIP Communica-
`tions For Dummies, Avaya Custom Edition, 2006, Wiley Publishing,
`
`Inc" HObOken’ NJ’ USA'
`* cited by examiner
`.
`.
`.
`Prtmary ExamtneriDamel A Hess
`Assistant ExamineriLaura Gudorf
`(74) Attorney, Agent, or Firmilntellectual Property Law
`Office of Thomas Chuang
`
`ABSTRACT
`(57)
`A headset includes a finger pad on an exterior of the headset
`on which a finger ofa headset wearer is placed. The headset
`includes an optical line scanner which scans the finger pad
`and outputs a series of successive images of the finger placed
`on the finger pad. A headset processor processes the output of
`the optical line scanner to detect relative motion ofthe finger
`on the finger pad or detect tapping of the finger on the finger
`pad.
`
`20 Claims, 5 Drawing Sheets
`
`APPLE 1008
`
` 1
`
`1
`
`APPLE 1008
`
`

`

`US. Patent
`
`Dec. 15, 2009
`
`Sheet 1 of5
`
`US 7,631,811 B1
`
`
`
`FIG. 1
`
`2
`
`

`

`US. Patent
`
`Dec. 15, 2009
`
`Sheet 2 of5
`
`US 7,631,811 B1
`
`Power Source
`
`11
`
`Memory
`12
`
`Microphone
`14
`
`Speaker
`16
`
`User Interface 18
`
`Line
`
`Scanner 20
`
`2 \
`
`Processor
`10
`
`50
`
`FIG. 2
`
`Input Keys 6
`
`Light Source 22
`
`.
`Optical Sensor 26
`
`Finger Pad 4
`
`3
`
`

`

`US. Patent
`
`Dec. 15, 2009
`
`Sheet 3 of5
`
`US 7,631,811 B1
`
`
`
`FIG. 3
`
`4
`
`

`

`U.S. Patent
`
`Dec. 15, 2009
`
`Sheet 4 of 5
`
`US 7,631,811 B1
`
`Line Scanner for Scrolling
`
`FIG. 4A
`
`FIG. 48
`
`32
`
`mmw_
`
`34
`
`Vix
`
`FIG. 4C
`
`FIG. 4D
`
`5
`
`
`

`

`US. Patent
`
`Dec. 15, 2009
`
`Sheet 5 015
`
`US 7,631,811 B1
`
`Line Scanner for Selecting
`
`4
`
`4
`
`4
`
`Figure 5A
`
`36
`
`Figure SB
`
`Figure 50
`
`6
`
`6
`
`

`

`US 7,631,811 B1
`
`1
`OPTICAL HEADSET USER INTERFACE
`
`BACKGROUND OF THE INVENTION
`
`Recent developments in the telecommunications industries
`have produced telecommunications devices with increased
`capabilities. As a result, the complexity of interacting with
`these devices has increased. Headsets are now capable of
`doing more than being simple peripherals to legacy phones.
`For example, the headsets may control navigation through
`menus or files.
`However, headset form factors do not lend themselves well
`to traditional user interface technologies like keypads and
`displays which are suited for complex user man-machine
`interface interactions. For example, the available space on the
`headset housing is limited. In the prior art, headset user inter-
`faces typically consist of a small number of multifunction
`buttons and a multifunction visual indicator. This limited user
`
`interface makes access to more complex features and capa-
`bilities difficult and non-intuitive, particularly when the head-
`set is being worn. Visual indicators have limited use while the
`headset is being worn. Multifunction buttons are non-intui-
`tive and awkward to use.
`
`they offer
`As headsets become more “intelligent”,
`advanced features and functionality. With increased features
`and functionality, these headsets require more complex user
`interfaces. However, the limited physical size of headset
`housings makes it desirable to minimize the number of or
`required size of the headset user interface mechanisms.
`As a result, there is a need for improved methods and
`apparatuses for headset user interface input mechanisms.
`
`BRIEF DESCRIPTION OF THE DRAWINGS
`
`The present invention will be readily understood by the
`following detailed description in conjunction with the accom-
`panying drawings, wherein like reference numerals designate
`like structural elements.
`
`FIG. 1 illustrates a headset capable ofreceiving user inputs
`utilizing an optical line scanner.
`FIG. 2 illustrates a simplified block diagram of the com-
`ponents of the headset shown in FIG. 1.
`FIG. 3 illustrates a side view of a headset showing the
`internal arrangement of a line scanner system.
`FIGS. 4A-4D illustrate sample operation of a line scanner
`to detect user input scrolling as the user “wipes” his finger.
`FIGS. 5A-5C illustrate sample operation of a line scanner
`to detect user input tapping.
`
`DESCRIPTION OF SPECIFIC EMBODIMENTS
`
`Methods and apparatuses for a headset user interface is
`disclosed. The following description is presented to enable
`any person skilled in the art to make and use the invention.
`Descriptions of specific embodiments and applications are
`provided only as examples and various modifications will be
`readily apparent to those skilled in the art. The general prin-
`ciples defined herein may be applied to other embodiments
`and applications without departing from the spirit and scope
`ofthe invention. Thus, the present invention is to be accorded
`the widest scope encompassing numerous alternatives, modi-
`fications and equivalents consistent with the principles and
`features disclosed herein. For purpose of clarity, details relat-
`ing to technical material that is known in the technical fields
`related to the invention have not been described in detail so as
`
`not to unnecessarily obscure the present invention.
`
`5
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`40
`
`45
`
`50
`
`55
`
`60
`
`65
`
`2
`
`This invention relates generally to the field of headset user
`interfaces and specifically to the field ofheadset user interface
`input mechanisms. In one example, this description describes
`a method and apparatus for a headset with an optical line
`scanner on a lightweight headset, where the optical line scan-
`ner detects finger movements, such as tapping, sliding for-
`ward and sliding backward to be translated into various
`inputs, such as volume up and down, menu scrolling, and
`other headset user interface options known in the art.
`As a user moves his or her finger moves across the line
`scanner, relative motion of the fingerprints ridges and valleys
`are scanned. In the same manner that an optical mouse inter-
`prets the changing images to detect movement of the mouse,
`motion of the user finger is determined. However, for a head-
`set user interface, only one axis is needed for scrolling, allow-
`ing the possibility of using a line scanner. For tapping, an
`algorithm is used to determine the amount of light being
`received by the optoelectronic sensor. In a further example, to
`reduce false triggers, such as due to hair falling in front ofthe
`sensor, secondary mechanisms are used such as overlaying a
`transparent touch sensor such as a capacitance sensor on the
`line scanner pad.
`In one example of the invention, a headset includes a
`microphone, a speaker, and a finger pad on an exterior of the
`headset on which a finger of a headset wearer is placed. The
`headset includes an optical line scanner which scans the fin-
`ger pad and outputs a series of successive images ofthe finger
`placed on the finger pad. A headset processor processes the
`output of the optical line scanner to detect relative motion of
`the finger on the finger pad or detect tapping of the finger on
`the finger pad. The optical line scanner may include a light
`source, an optical guide for forming a line of light from the
`light source, an imaging sensor, and a lens for directing the
`line of light reflected from the finger pad onto the imaging
`sensor.
`
`In one example ofthe invention, a headset includes a finger
`receiving means for placement of a user finger, and an optical
`line scanning means for scanning the finger receiving means
`on a headset housing exterior and providing an output of
`successive images ofthe finger receiving means. The process-
`ing means processes the output of successive images on the
`finger pad to determine a relative movement of a user finger
`across the finger pad or to determine a tapping of the user
`finger on the finger pad. The processing means modifies a
`headset control operation responsive to the relative move-
`ment of the user finger or the tapping.
`In one example of the invention, a method for receiving
`user input at a headset includes providing a transparent finger
`pad on a headset housing for receiving a user finger, providing
`an optical line scanner disposed within the headset housing,
`and scanning the transparent finger pad with the optical line
`scanner to output a series of successive images. The method
`further includes processing the successive images to deter-
`mine a relative movement of a user finger across the finger
`pad, and modifying a headset control operation responsive to
`the relative movement of the user finger.
`In one example ofthe invention, a headset includes a finger
`receiving means for placement of a user finger, and an optical
`scanning means for scanning the finger receiving means on a
`headset housing exterior and providing a sequence of electri-
`cal signals associated with successive line scan images of the
`finger receiving means. The headset further includes a pro-
`cessing means for processing the sequence of electrical sig-
`nals to determine a relative bi-directional movement of a user
`
`finger across the finger pad along a single axis or determine a
`tapping of the user finger on the finger pad. The processing
`
`7
`
`

`

`US 7,631,811B1
`
`3
`means modifies a headset control operation responsive to the
`relative movement of the user finger or the tapping.
`In one example of the invention, a method for receiving
`user input at a headset includes providing a transparent finger
`pad on a headset housing for receiving a user finger, providing
`an optical line scanner disposed within the headset housing,
`and scanning the transparent finger pad with the optical line
`scanner to output a sequence of electrical signals associated
`with successive line scan images. The sequence of electrical
`signals is processed to determine a relative movement of a
`user finger across the finger pad. The headset control opera-
`tion is modified responsive to the relative movement of the
`user finger.
`FIG. 1 illustrates a headset 2 capable of receiving user
`inputs utilizing an optical line scanner. Headset 2 includes a
`narrow finger pad 4 serving as a scanning surface on which a
`user finger is placed and scanned by the user “wiping” his
`finger across the scanning surface. During optical scanning,
`the user slides his or her finger across the scanning surface,
`whereby the line scanner images the finger line by line as it is
`slid across the scanning surface.
`FIG. 2 illustrates a simplified block diagram of the com-
`ponents of the headset 2 shown in FIG. 1. The headset 2
`includes a processor 10 operably coupled via a bus 50 to a
`memory 12, a microphone 14, power source 11, and user
`interface 18. User interface 18 includes a line scanner 20 and,
`optionally, one or more input buttons or keys 6. In one
`example, line scanner 20 includes a light source 22, lens 24,
`and optical sensor 26. Optical sensor 26 is, for example, a
`charge coupled device (CCD) such as a CMOS square pixel
`array. The CCD is an array of light sensitive diodes which
`generate an electrical signal in response to light which hits a
`particular pixel. Line scanner 20 may also include a processor
`for processing scan data. Alternatively, line scanner 20 may
`utilize processor 10 to process scan data. Line scanner 20 may
`also include memory separate from memory 12 for storing
`scan data or firmware/software executable to operate line
`scanner 20 and process scan data. The firmware/software may
`include a user input
`identifier application for analyzing
`scanned finger motion data to determine user input at the
`finger pad 4. Alternatively,
`line scanner 20 may utilize
`memory 12 for such purposes. The line scanner 20 is properly
`aligned and integrated with finger pad 4 within the headset
`housing. In a further example, line scanner 20 is replaced with
`an alternative optical scanner. Examples of optical scanners
`include, without limitation, image sensors, planar scanners,
`CMOS sensors, contact image sensors, or other optical sys-
`tems such as used by optical mouse devices.
`Memory 12 may include a variety of memories, and in one
`example includes SDRM, ROM, flash memory, or a combi-
`nation thereof. Memory 12 may further include separate
`memory structures or a single integrated memory structure. In
`one example, memory 12 may be used to store passwords,
`network and telecommunications programs, and/or an oper-
`ating system (OS).
`Processor 10, using executable code and applications
`stored in memory, performs the necessary functions associ-
`ated with headset operation described herein. Processor 10
`allows for processing data,
`in particular managing data
`between user interface 18 and operation of headset 2 func-
`tions. In one example, processor 10 is a high performance,
`highly integrated, and highly flexible system-on-chip (SOC),
`including signal processing functionality such as echo can-
`cellation/reduction and gain control in another example. Pro-
`cessor 10 may include a variety of processors (e.g., digital
`signal processors), with conventional CPUs being applicable.
`
`5
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`40
`
`45
`
`50
`
`55
`
`60
`
`65
`
`4
`
`In one example the line scanner 20 continuously monitors
`finger pad 4 to identify whether a user has placed a finger on
`finger pad 4 to perform an input action. User interface 18
`allows for manual communication between the headset user
`
`and the headset. User interface 18 may also include, for
`example, an audio and/or visual interface such that an audio
`prompt may be provided to the user’s ear and/or an LED may
`be lit. For example, the prompt may inform the user to place
`his or her finger on the finger pad to perform a fingerprint
`scan. Although illustrated in FIG. 2 as separate from line
`scanner 20, finger pad 4 may also be considered to be a
`component of line scanner 20.
`FIG. 3 illustrates a side view of the headset 2 showing the
`internal arrangement of the line scanner system. The housing
`body ofheadset 2 includes a finger pad 4. Headset 2 may also
`include one or more user interface buttons or keys 6 which the
`user may depress.
`The finger pad 4 is optically transparent, allowing light
`from a light source 22 disposed within the headset housing to
`exit the headset. For example, finger pad 4 is a planar surface
`composed of glass or plastic. After light from the light source
`22 is reflected off the finger pad 4, it re-enters the headset
`housing and is focused by a lens 24 on an optical sensor 26.
`For example, optical sensor 26 is a motion sensor integrated
`circuit (IC) having an array ofphotodetectors for capturing an
`image. Optical sensor 26 converts light information into an
`electrical signal and transmits the signal to a system image
`processing unit. Lens 24 and optical motion sensor 26 are
`disposed within the headset 2. An optical guide may be
`arranged to convert light from light source 22 into a line-
`shaped light to illuminate a fingerprint in a line-shape. In a
`further example, a scanning component may sweep a beam
`spot of light across the finger pad 4 and the optical sensor may
`be a single element.
`The light reflected off the finger pad 4 forms an image of a
`user finger placed on the finger pad 4 on optical sensor 26.
`This image is captured by optical sensor 26. For example,
`optical sensor 26 has a plurality of line-shaped photoelectric
`converting elements. Successive images of a finger placed on
`the finger pad 4 are then compared by a processor. The pro-
`cessor may be integrated with the optical sensor 26 or may be
`a separate processor such as the headset processor 10. The
`successive images are compared to determine the forward or
`backward motion of the user finger across finger pad 4. The
`successive images are also compared to determine whether
`the user is “tapping” or “double tapping” the finger pad 4, i.e.,
`quickly placing his finger tip on finger pad 4 and then remov-
`ing it. Depending on the current operational state ofthe head-
`set, the forward or backward motion is translated to a pre-
`defined user input, such as scrolling through a menu or
`volume increase or decrease. User tapping or double tapping
`is translated, for example, to a user selected command.
`The directional motion of a finger on finger pad 4 along a
`single axis (e.g., X or Y) or the presence of a finger on finger
`pad 4 is detected optically by optical sensor 26 by directly
`imaging, as an array of pixels, the various particular ridges
`and valleys of the user fingerprint placed on the finger pad 4.
`The particular features of the fingerprint are illuminated by
`the light source 22. The use of optical sensors to detect direc-
`tion and degree ofmovement along an X-Y coordinate system
`is described in US. Pat. No. 6,233,368 issued May 15, 2001,
`entitled “CMOS Digital Optical Navigation Chip”, which is
`hereby incorporated by reference for all purposes.
`The motion of a finger on finger pad 4 is detected by optical
`sensor 26 by comparing a newly captured image with a pre-
`viously captured image to ascertain the direction and amount
`ofmovement. The newly captured image and previously cap-
`
`8
`
`

`

`US 7,631,811 B1
`
`10
`
`5
`tured images may be stored in memory on the optical sensor
`26 or stored in headset memory 12. For example, referring to
`FIGS. 4A-4C, sample operation of line scanner 20 to detect
`user input scrolling as the user “wipes” his finger in a direc-
`tion 30 is illustrated. FIG. 4A illustrates a fingerprint 28 5
`corresponding to a user finger placed on finger pad 4. Refer-
`ring to FIG. 4C, a fingerprint portion 32 scanned by line
`scanner 20 is captured. At an immediate point in time there-
`after as the user finger is wiped in direction 30, the fingerprint
`28 is at a second position on finger pad 4, as shown in FIG. 4B.
`As a result, a different fingerprint portion 34 scanned by line
`scanner 20 is captured, as shown in FIG. FIG. 4D. The image
`of fingerprint portion 32 is compared to the image of finger-
`print portion 34 to ascertain the direction of movement of the
`user finger across finger pad 4. For example, the ridges and 15
`valleys of the fingerprint line scans may be pattern matched
`and aligned to determine the direction of movement.
`For example, referring to FIGS. 5A-5C, sample operation
`of line scanner 20 to detect user input tapping is illustrated. In
`this example, line scanner 20 detects user tapping by deter- 20
`mining whether the user has placed his finger across finger
`pad 4 and removed it immediately thereafter. For example, at
`a first time shown in FIG. 5A, the user finger pad 4 does not
`have a finger placed upon it. At a second time shown in FIG.
`5B, the user has placed his finger upon finger pad 4, which is 25
`scanned by line scanner 20. At third time illustrated in FIG.
`5C, the finger pad 4 is once again clear as the user has
`removed his finger. In one example, the quantity of light
`detected by the sensor is used to determine a tap. In a further
`example, a time period on which the user finger is placed on 30
`finger pad 4 to indicate a user tap is empirically determined.
`Double tapping is detected, for example, by detecting user
`tapping twice within a predefined time period.
`The information developed by optical sensor 26 regarding
`the motion of the user finger on the finger pad 4 is relayed to 35
`the headset processor 10, which translates the information to
`correspond to user input actions at the headset. The headset
`processor 10 then implements the desired input action. For
`example, such desired input actions may include volume
`control, power control, call answer, call
`terminate,
`item 40
`select, next item, and previous item, or other actions typically
`performed at a headset device.
`In a further example, line scanner 20 is used to authenticate
`the identity of the headset user by scanning the fingerprint of
`the user and comparing it to a previously stored authorized 45
`fingerprint. During optical scanning, the user slides his or her
`finger across the scanning surface, whereby the line scanner
`images the finger line by line as it is slid across the scanning
`surface. In this manner, the fingerprint of the user is gener-
`ated. In this example, the headset memory includes previ- 50
`ously stored fingerprint data corresponding to validated users,
`a feature identifier application for analyzing scanned finger-
`print scan data, and a fingerprint match application for com-
`paring the analyzed scanned fingerprint scan data to previ-
`ously stored fingerprint data. In one example, headset user 55
`authentication is required prior to allowing the user to operate
`the headset. In this example, the line scanner 20 serves the
`dual function of being a user interface input device and an
`authentication device.
`
`The various examples described above are provided by 60
`way of illustration only and should not be construed to limit
`the invention. Based on the above discussion and illustrations,
`those skilled in the art will readily recognize that various
`modifications and changes may be made to the present inven-
`tion without strictly following the exemplary embodiments 65
`and applications
`illustrated and described herein. For
`example, the methods and systems described herein may be
`
`6
`applied to other body worn devices in addition to headsets.
`Furthermore, the functionality associated with any blocks
`described above may be centralized or distributed. It is also
`understood that one or more blocks of the headset may be
`performed by hardware, firmware or software, or some com-
`binations thereof. Such modifications and changes do not
`depart from the true spirit and scope of the present invention
`that is set forth in the following claims.
`While the exemplary embodiments ofthe present invention
`are described and illustrated herein, it will be appreciated that
`they are merely illustrative and that modifications can be
`made to these embodiments without departing from the spirit
`and scope of the invention. Thus, the scope of the invention is
`intended to be defined only in terms ofthe following claims as
`may be amended, with each claim being expressly incorpo-
`rated into this Description of Specific Embodiments as an
`embodiment of the invention.
`
`What is claimed is:
`
`1. A headset comprising:
`a microphone;
`a speaker;
`a finger pad on an exterior of the headset on which a finger
`of a headset wearer is placed;
`an optical line scanner, wherein the optical line scanner
`scans the finger pad and outputs a series of successive
`images of the finger placed on the finger pad; and
`a processor, wherein the processor processes the series of
`successive images to detect relative motion of the finger
`on the finger pad or detect tapping of the finger on the
`finger pad.
`2. The headset of claim 1, wherein the optical line scanner
`comprises:
`a light source;
`an optical guide for forming a line of light from the light
`source an imaging sensor; and
`a lens for directing the line of light reflected from the finger
`pad onto the imaging sensor.
`3. The headset of claim 2, wherein the light source com-
`prises a light emitting diode.
`4. The headset of claim 2, wherein the imaging sensor
`comprises an integrated circuit sensor.
`5. The headset of claim 1, wherein the optical line scanner
`detects relative motion of the finger on the finger pad along a
`single axis.
`6. The headset of claim 1, wherein the fingerpad comprises
`a glass planar surface.
`7. A headset comprising:
`a finger receiving means for placement of a user finger;
`an optical line scanning means for scanning the finger
`receiving means on a headset housing exterior and pro-
`viding an output of successive images of the finger
`receiving means;
`a capacitive sensing means overlaid on the finger receiving
`means for sensing placement of a user finger to detect a
`false trigger of the optical line scanning means; and
`a processing means for processing the output of successive
`images of the finger receiving means to determine a
`relative movement of a user finger across the finger
`receiving means or determine a tapping ofthe user finger
`on the finger receiving means, wherein the processing
`means modifies a headset control operation responsive
`to the relative movement of the user finger or the tap-
`ping;
`a first transducer means for receiving a user speech signal;
`and
`
`a second transducer means for outputting an audio signal.
`
`9
`
`

`

`US 7,631,811 B1
`
`7
`8. The headset of claim 7, wherein the headset control
`operation comprises one or more selected from the following
`group: volume control, power control, call answer, call ter-
`minate, item select, next item, and previous item.
`9. A method for receiving user input at a headset compris-
`ing:
`providing a transparent finger pad on a headset housing for
`receiving a user finger;
`providing an optical line scanner disposed within the head-
`set housing;
`scanning the transparent finger pad with the optical line
`scanner to output a series of successive images;
`processing the series of successive images to determine a
`relative movement of a user finger across the transparent
`finger pad; and
`modifying a headset control operation responsive to the
`relative movement of the user finger.
`10. The method of claim 9, wherein processing the succes-
`sive images to determine a relative movement of a user finger
`across the finger pad comprising determining whether the
`user finger is moving in a first direction along an axis or in a
`second direction opposite the first direction along the axis.
`11. The method of claim 9, further comprising processing
`the successive images to determine a tap of the user finger on
`the finger pad.
`12. The method of claim 11, wherein processing the suc-
`cessive images to determine a tap of the user finger on the
`finger pad comprises measuring a quantity of light received at
`the optical line scanner.
`13. The method of claim 9, wherein processing the succes-
`sive images comprises identifying and comparing fingerprint
`ridges and valleys.
`14. A headset comprising:
`a finger receiving means for placement of a user finger;
`an optical scanning means for scanning the finger receiving
`means on a headset housing exterior and providing a
`sequence of electrical signals associated with successive
`line scan images of the finger receiving means;
`a capacitive sensing means overlaid on the finger receiving
`means for sensing placement of a user finger to detect a
`false trigger of the optical line scanning means; and
`a processing means for processing the sequence of electri-
`cal signals to determine a relative bi-directional move-
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`40
`
`8
`ment of a user finger across the finger receiving means
`along a single axis or determine a tapping of the user
`finger on the finger receiving means, wherein the pro-
`cessing means modifies a headset control operation
`responsive to the relative bi-directional movement ofthe
`user finger or the tapping;
`a first transducer means for receiving user speech; and
`a second transducer means for outputting an audio signal.
`15. The headset of claim 14, wherein the headset control
`operation comprises one or more selected from the following
`group: volume control, power control, call answer, call ter-
`minate, item select, next item, and previous item.
`16. A method for receiving user input at a headset compris-
`ing:
`providing a transparent finger pad on a headset housing for
`receiving a user finger;
`providing an optical line scanner disposed within the head-
`set housing;
`scanning the transparent finger pad with the optical line
`scanner to output a sequence of electrical signals asso-
`ciated with successive line scan images;
`processing the sequence of electrical signals to determine a
`relative movement of a user finger across the transparent
`finger pad; and
`modifying a headset control operation responsive to the
`relative movement of the user finger.
`17. The method of claim 16, wherein processing the
`sequence of electrical signals comprises pattern matching
`fingerprint ridges and valleys.
`18. The method of claim 16, wherein processing the
`sequence of electrical signals to determine a relative move-
`ment of a user finger across the finger pad comprising deter-
`mining whether the user finger is moving in a first direction
`along an axis or in a second direction opposite the first direc-
`tion along the axis.
`19. The method of claim 16, further comprising processing
`the sequence ofelectrical signals to determine a tap ofthe user
`finger on the finger pad.
`20. The method of claim 19, wherein processing the
`sequence of electrical signals to determine a tap of the user
`finger on the finger pad comprises measuring a quantity of
`light received at the optical line scanner.
`*
`*
`*
`*
`*
`
`10
`
`10
`
`

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket