throbber
as) United States
`a2) Patent Application Publication (0) Pub. No.: US 2007/0078340 Al
`(43) Pub. Date: Apr. 5, 2007
`
`Wilcox et al.
`
`US 20070078340A1
`
`(54) METHOD AND APPARATUS FOR
`CONTROLLING ULTRASOUND IMAGING
`SYSTEMS HAVING POSITIONABLE
`TRANSDUCERS
`
`(22)
`
`Filed:
`
`Sep. 30, 2005
`
`Publication Classification
`
`(75)
`
`Inventors: Stephen D. Wilcox, Los Gatos, CA
`(US); Sankaralingam Ramraj,
`Sunnyvale, CA (US)
`
`Correspondence Address:
`STEMENS CORPORATION
`INTELLECTUAL PROPERTY DEPARTMENT
`170 WOOD AVENUE SOUTH
`
`ISELIN, NJ 08830 (US)
`
`(73) Assignee: Siemens Medical Solutions USA, Inc.
`
`(21) Appl. No.:
`
`—-11/242,161
`
`(51)
`
`Int. Cl.
`(2006.01)
`A6GIB 8/00
`(52) US. C1. eee cecssseceseessesceneneesssecsssensssenseneeneaee 600/437
`
`(57)
`
`ABSTRACT
`
`A method and system for providing an operational command
`signal to a workstation of an imaging system. The worksta-
`tion is provided imagingdata from a positionable transducer.
`The method and system convert at least one of a predeter-
`mined plurality of motion patterns imparted by an operator
`of the system to the transducer into the operational com-
`mandsignal.
`
`14
`
`

`

`Patent Application Publication Apr.5,2007 Sheet 1 of 12
`
`US 2007/0078340 Al
`
`

`

`Patent Application Publication Apr.5,2007 Sheet 2 of 12
`
`US 2007/0078340 Al
`
`15
`
`Lv.
`
`FIG. 2A
`
`20
`
`
`
`FIG. 2B
`
`

`

`US 2007/0078340 Al
`
`Patent Application Publication Apr.5,2007 Sheet 3 of 12
`
`FIG. 2C
`
`20
`
`
`
`
`CNT
`
`
`NTT
`
`TTT
`
`
`OUDDEUOREORREOOREONOOEROREROREOUELL
`OUUDEUOARAROOGROOOROERGOREGOOROUERH
`
`
`UT
`
`
`OUDEEUOARARUGROGGRREOREORROOOEOUERI
`TEEEE
`
`
`FIG. 2D
`
`

`

`Patent Application Publication Apr.5, 2007 Sheet 4 of 12
`
`US 2007/0078340 Al
`
`obeu
`
`JOSSe00i¢
`
`weeg
`
`Burwi0o4
`
`ueoS
`
`JaYaAU0D
`
`oyoz
`
`JOSSED0Jq
`
`OMION
`
`

`

`Patent Application Publication Apr.5,2007 Sheet 5 of 12
`
`US 2007/0078340 Al
`
`aD Transducer
`
`X Axis (Elevation)
`"Forward/
`
`Backward"
`
`1D Array,
`2D Image
`
`"U p"
`
`20
`
`Transducer Face
`a’
`
`"Right"
`
`Y Axis (Azimuth)
`
`"Down"
`
`FIG. 4A
`
`
`
`ZAxis(Axial)
`
`
`
`
`
`15
`
`2D Transducer
`
`12
`
`X Axis (Elevation)
`
`individual Transducer
`Elemets, 19
`
`Transducer Face
`
`
`Y Axis (Azimuth)
`
`2D Array,
`3D Image
`
`
`
`ZAxis(Axial)
`
`FIG. 4B
`
`

`

`(Axis me"?
`AyOrn Vy
`<—(f XY j )
`i /
`\
`|
`py!
`\
`\
`/
`
`/
`
`/
`/
`/
`/
`!
`/
`
`FIG. SA
`
`Patent Application Publication Apr.5,2007 Sheet 6 of 12
`
`US 2007/0078340 Al
`
`12
`
`—> “7
`
`\
`
`!
`/
`
`/
`
`/
`
`/
`
`FIG. 5C }
`
`i
`
`

`

`Patent Application Publication Apr.5,2007 Sheet 7 of 12
`
`US 2007/0078340 Al
`
`Pictograph——
`
`
`
`
`
`|
`Y
`z
`12
`
`
`
`Userinterface action
`ae
`
`
`len|<a
`F SS
`R-L-R-L
`Ys Mouse Rightclick, bring up a menu.
`
`
`
`
`WN
`AL-R-R-L
`
`
`
`
`
`
`7
`
`
`jo|@|apSS
`
`
`
`
`
`
`
`
`
`
`
`
`13)R-L-D-U
`
`GZ
`
`_R./P.
`
`
`
`
`Goto the next stage in a defined exam protocol. This may
`change a combination of imaging parameters, stopwatch
`timers, image annotations, measurementtools and calculation
`33 package measurements.
`
`Display the next entry in a series of pre-defined
`
`
`
`Explanation of pictographs:
`
`
`
`
`30_f\= Vector Image
`
`
`= Path of transducer motion
`
`
`
`
`
`
`
`
`FIG. 6A
`
`
`
`
`
`
`
`
`of measurements.
`it Mouseleft click.
`
`SS
`
`9S
`
`S
`
`iW
`NJ
`a=
`
`Toyo[=EPIERIE
`
`R-L-L-R
`
`U-D
`
`
`
`
`
`
`
`
`
`

`

`Patent Application Publication Apr.5,2007 Sheet 8 of 12
`
`US 2007/0078340 Al
`
`Y Axis Position
`
`Triggerlevels resetif pattern
`not detected within
`timout period
`
`Sonographer
`command
`signal motion
`
`-— ee
`
`
`
`Motion that
`occurs during
`normal scanning
`
`
`
`
`
`Timer Timout
`
`Time
`
`Timerstarted at
`first level crossing
`
`FIG. 6B
`
`

`

`Patent Application Publication Apr.5, 2007 Sheet 9 of 12
`
`US 2007/0078340 Al
`
`Scanneracquires image data
`
`Determine overall image movement vector
`
`
`
`
`Yes
`
`
`Motion
`
`greater than
`threshold?
`
`
`Acquire additional image data
`
`
`Determine additional overall
`image motion vector
`
`threshold?
`
`
`
`
`Matches
`
`a knownpattern
`
`template?
`
`
`
`
`
` Motion
`greater than
`
`Yes
`
`Compare motion vector pattern
`to pattern templates
`
`
`LookupUserInterface Action for pattern
`
`Yes
`
`Send commandto userinterface software
`
`700
`
`702
`
`704
`
`706
`
`708
`
`710
`
`712
`
`FIG. 7
`
`

`

`Patent Application Publication Apr.5,2007 Sheet 10 of 12
`
`US 2007/0078340 Al
`
`
`Yes
`
`
`
`
`
`
` Motion
`
`
`greater than
`threshold?
`
`
`
`
`Timer
`
`
`not expired?
`
`Yes
`
`Acquire additional image data
`
`Determine additional overall image motion vector
`
` Motion
`
`
`
`greater than
`hreshold?
`
`
`Yes
`
`To Fig. 8B
`
`FIG. 8A
`
`810
`
`}—812
`
`

`

`Patent Application Publication Apr.5,2007 Sheet 11 of 12
`
`US 2007/0078340 Al
`
`FIG. 8
`
`FIG. 8A
`
`FIG. 8B
`
`From Fig. 8A
`
`
`
` Timer
`not expired?
`
`Yes
`
`Compare motion vectorpattern to
`pattern templates
`
`
`
`
`Matches
`a knownpattern
`template?
`
`Yes
`
`
`
`814
`
`816
`
`818
`
`
`
`
`
`
`Lookup UserInterface Action for pattern
`
`
`Send command to userinterface software
`
`FIG. 8B
`
`

`

`Patent Application Publication Apr.5, 2007 Sheet 12 of 12
`
`US 2007/0078340 Al
`
`image
`
`Sub-image
`segments
`
`Sub-image
`LZ.
`segments used
`detect overall
`image motion
`
`Sub-
`
`image
`
`a&&@7)
`
`oOwosBw
`
`detect overall
`image motion
`
`
`

`

`US 2007/0078340 Al
`
`Apr. 5, 2007
`
`METHOD AND APPARATUS FOR CONTROLLING
`ULTRASOUND IMAGING SYSTEMS HAVING
`POSITIONABLE TRANSDUCERS
`
`
`
`TECHNICAL FIELD
`
`[0001] This invention relates generally to methods and
`apparatus for controlling imaging systems and more par-
`ticularly for controlling imaging systems having position-
`able transducers.
`
`BACKGROUND
`
`Asis knowninthe art, one type of imaging system
`[0002]
`is an ultrasound imaging systems. A conventional ultrasound
`imaging system includesa positional transducer, typically a
`sonographer handheld transducer, coupled to a large pro-
`cessing and display workstation or operator interface. The
`frontal portion of the transducer includes an array ofultra-
`sonic elements which transmit and receive ultrasonic energy
`for imaging a selected region of a patient. The received
`ultrasonic energy is converted to electric signals by the
`transducer and passed to the workstation. The workstation
`detects, filters and otherwise processes the information to
`generate a two- or three-dimensional representation of the
`scanned region.
`
`[0003] The sonographer supplies the control signals for
`the workstation. Such control signals are typically supplied
`by the sonographer’s free, or non-transducer carrying hand.
`Scanning situations in both the examination room and other
`locations often require the sonographer to be in awkward
`positions for simultaneously reaching the controls with the
`free hand and placing the frontal portion of the transducer in
`the proper position on the patient’s body. One technique
`suggested to solve this problem is through voice activation;
`however, such technique may be proneto error and requires
`a speech recognition learning phase for each sonographer.
`Another technique suggested to provide the central signals
`to the workstation is through a sonographer actuated foot
`pedal; however such is not practical for all scanning situa-
`tions.
`
`SUMMARY
`
`In accordance with the present invention, a method
`[0004]
`is provided for providing an operational commandsignals,
`sometimes herein referred to as control signals, to a work-
`station of an imaging system. The workstation is provided
`imaging data from a positionable transducer. One method
`includes converting at least one of a plurality of predeter-
`mined motion patterns imparted by an operatorof the system
`to the transducer into the operational commandsignals.
`
`[0005] With such method, the operator is provided with a
`wayto control the workstation without taking a handoff of
`the transducer, or relying on voice control or foot actuated
`controls. The method reduces the number of times the
`operator must touch controls on the workstation.
`
`[0006] Another method includes parting the transducer
`head into multiple regions (in one embodiment, less than
`four regions) and interpreting the reception of the signals
`from such regions into operational commandsignals.
`
`the converting comprises
`In one embodiment,
`[0008]
`detecting at least one of the predetermined motion patterns
`and converting such detected motion patterns into a corre-
`sponding one of the operational commandsignals.
`
`In one embodiment, the detecting comprises com-
`[0009]
`paring a sequence of images formed by the system.
`
`Inone embodiment, the methodincludes determin-
`[0010]
`ing from the sequence of images whether the motion
`imparted to the transducer waseither a repositioning of the
`transducer to producea different image to be observed by the
`operator or a motion imparted to produce the corresponding
`one of the commandsignals to the workstation.
`
`In one embodiment, such determining includes
`[0011]
`comparing types of motions imparted bythe operation.
`
`In one embodiment, such determining includes
`[0012]
`comparing imparted motion with a level threshold.
`
`In one embodiment, such determining includes
`[0013]
`comparing imparted motion with a time duration threshold.
`
`In one embodiment, a method is provided for
`[0014]
`providing control signals to a workstation of an imaging
`system, such workstation being provided imaging data from
`a positionable transducer. The method includes detecting
`patterns of motion of the transducer, and converting the
`patterns to the control signals.
`
`In one embodiment, the detection is performed by
`[0015]
`detecting patterns of changein real time images provided by
`the system.
`
`In one embodiment, timing of the motion is used to
`[0016]
`discriminate between motion intendedto provide the control
`signals and motion normally occurring during scanning.
`
`In one embodiment, patterns of direction of the
`[0017]
`transducer motion are used to discriminate between motion
`
`intended to provide the control signals and motion normally
`occurring, during scanning.
`
`In one embodiment, a combination of patterns of
`[0018]
`direction of the transducer motion and timing of the motion
`are used to discriminate between motion intendedto provide
`the control signals and motion normally occurring during
`scanning.
`
`In one embodiment, an imaging system is provided
`[0019]
`having a workstation and a positionable transducer for
`providing imaging data to the workstation. The workstation
`responds to the operational commandsignals. The worksta-
`tion includes a memory for storing a table mapping detected
`motion of the transducer into the commandsignals.
`
`the workstation includes a
`In one embodiment,
`[0020]
`processor programmedto detect at least one of a predeter-
`minedplurality of motion patterns and convert such detected
`one of the motion patterns into the operational command
`signals.
`
`In one embodiment, the transducer has deposed
`[0021]
`within a housing thereof motion sensors.
`
`In one embodiment, sensors disposed remote from
`[0022]
`the transducer sense motion of the transducer.
`
`[0007] A third method includes converting detections of
`predetermined echo signatures into operational command
`signals.
`
`In one embodiment, an imaging system is provided
`[0023]
`having a workstation and a positionable transducer for
`providing imaging data to the workstation. The workstation
`
`

`

`US 2007/0078340 Al
`
`Apr. 5, 2007
`
`responds to control signals. The workstation includes a
`processor for detecting patterns of motion of the transducer
`and converting the patterns to the control signals.
`
`[0024] The details of one or more embodiments of the
`invention are set forth in the accompanying drawings and
`the description below. Other features, objects, and advan-
`tages of the invention will be apparent from the description
`and drawings, and from the claims.
`
`DESCRIPTION OF DRAWINGS
`
`FIG.1 is a sketch of an imaging system according
`[0025]
`to the invention;
`
`[0026] FIG. 2A is an isometric sketch of a transducer used
`in the system of FIG. 1;
`
`[0027] FIG. 2B is a front elevation view of the frontal
`portion of the transducer of FIG. 2A;
`
`[0028] FIG. 2C is an isometric sketch of a transducer
`adapted for use in the system of FIG. 1 according to one
`embodimentof the invention;
`
`[0029] FIG. 2D is a front elevation view of the frontal
`portion of a 2D array transducer;
`
`[0030] FIG. 3 is a block diagram of a processor used to
`process imaging data from the transducer of FIGS. 2A and
`2B to generate an image for an operator of a workstation
`used in the system of FIG. 1 and to process such data to
`generate control signal for operation of the workstation;
`
`[0031] FIG. 4A is a diagram showing a coordinate system
`for a one-dimensionalarray transducer used in the system of
`FIG. 1, such coordinate system indicating and defining
`operator motion of the transducer;
`
`FIG.4B is a diagram showing a coordinate system
`[0032]
`for a two-dimensional array transducer used in the system of
`FIG. 1, such coordinate system indicating and defining
`operator motion of the transducer;
`
`[0033] FIGS. 5A, 5B and 5C show various patterns of
`motion which may be imparted to the transducer by the
`operator and then image data which is processed by the
`processor of FIG. 3 to generate the control signally for the
`workstation of the system of FIG. 1;
`
`[0034] FIG. 6A is a table showing the relationship
`between a repertoire of motions impartable by the operator
`to the transducer and the workstation control
`signals
`intended by the operator resulting from such motions;
`
`[0035] FIG. 6B shows a comparison between normal
`transducer motion occurring during scanning and motion
`used to initiate a commandsignalto the workstation of FIG.
`1;
`
`[0036] FIG. 7 is a flow diagram of a process used by the
`processor of FIG. 3 in generating the workstation control
`signals from images generated by the transducer;
`
`[0037] FIGS. 8, 8A and 8B are flow diagrams in more
`detail of the process of FIG. 7 used by the processor of FIG.
`3 in generating the workstation control signals from images
`generated by the transducer;
`
`[0038] FIG. 9 is a diagram showing an image obtained by
`the processor of FIG. 3 divided into sub-image segments by
`the processor and used to detect overall image motion; and
`
`[0039] FIG. 10 is a diagram showing an imageobtained by
`the processor of FIG. 3 divided into sub-image segments by
`the processor with near-field segments used to detect overall
`image motion.
`
`[0040] Like reference symbols in the various drawings
`indicate like elements.
`
`DETAILED DESCRIPTION
`
`[0041] FIG. 1 shows an imaging system 10, here an
`ultrasound imaging system for medical diagnostics. The
`system 10 includes a positionable, here handheld,
`image
`processing ultrasound device, here a transducer 12 shown in
`more detail in FIGS. 2A and 2B, and a multi-use display
`device, or operator interface, herein sometime collectively
`referred to as a workstation 14. The handheld transducer 12
`obtains ultrasound data and formats the ultrasound data for
`transmission to the workstation 14, here via a cable 15.
`Controls to the workstation 14 are provided by detecting
`patterns of motion, to be described in more detail below,
`provided, or imparted, to the transducer 12 by an operator of
`the system 10, typically by a sonographer. Suffice it to say
`here that the pattern of motions may include: an up-down or
`down-up(i.e., axial) motion of the sonographer’s transducer
`hand holding wrist (i.e., an up flick of the wrist followed by
`a downflick of the wrist or a sequence of down-up flicks of
`the wrist); a left-rightor right-left (1.e., azimuthal) motion of
`the sonographer’s transducer hand holding wrist (i.e., left-
`rightflicks or right-left flicks of the wrist); for 2D arrays, a
`forward-and-backward (1.e., elevational) motion of the
`sonographer’s transducer hand holding wrist; a inward-
`outward motion towards and away from the patient’s body,
`or visa versa, or any combination thereof. The detection of
`these sonographer’s imparted transducer motions may be
`performed by hardware and/or software to detect patterns of
`change in real time images. When motion is detected, the
`timing of the motion is used to discriminate between motion
`intended to initiate control changes and motion which occurs
`normally during scanning. In addition, patterns of direction
`are be used to discriminate between motion intended to
`initiate control changes and motion which occurs normally
`during scanning. The combination of timing anddirection of
`transducer 12 motion changes are used to discriminate
`between transducer motion intended to initiate control
`
`changes and motion which occurs normally during scanning.
`
`[0042] The transducer 12 includes a housing 16 (FIG. 2A)
`adapted to be easily handheld, such as for example, being
`less than 8 inches in any dimension and/or having an
`ergonomic shape for holding in a operator’s hand. The
`housing 16 comprises plastic, rubber, metal, other materials
`now knownorlater developed, or combinations thereof. In
`one embodiment shownin FIG. 2A,the housing 16 is shaped
`for ergonomic use or holding by the operator(e.g., sonog-
`rapher) by having a generally round or curved circumference
`handle serving as a grip for the sonographer’s hand.
`
`[0043] The handheld transducer 12 includes conventional
`ultrasoundcircuitry, not shown, within the housing 16. Thus,
`the ultrasound circuitry includes, in the frontal portion 20
`thereof (FIGS. 2A and 2B) an array of ultrasonic elements
`19 which transmit and receive ultrasonic energy for imaging
`a patient, not shown. It
`is noted that FIG. 2B is for a
`one-dimensionalarray transducer and FIG. 2D is for a 2D
`array transducer. Here, the transducer’s transmit and receive
`
`

`

`US 2007/0078340 Al
`
`Apr. 5, 2007
`
`elements 19 in the frontal portion 20 are arranged in an
`elongated array along a the longeraxis, here the Y axis of the
`rectangular shaped patient interfacing surface, 1.e., the fron-
`tal portion 20, of the housing 16. The elements 20 in the
`frontal portion 20 are, as shown in FIG.3 fed to a display 22
`of the workstation 14 (FIG. 1) serially through: a beam-
`forming network 24, an echo processor 26, a scan converter
`28, and an image processor 31 in a conventional manner.
`The beamforming network 24, echo processor 26, scan
`converter 28, image processor 31 and display 22 are con-
`trolled by a central processing unit (CPU) 32 coupled to a
`random access memory RAM37. The CPU 32 operates in
`accordance with program instructions stored in a ROM 34,
`or in RAM 37,or in flash memory not shown, or on a hard
`drive device, not shown. A memory 36, here an erasable, or
`other type of programmable semiconductor memory, here a
`read only memory (ROM)is provided for storing a com-
`puter, here microprocessor, executable program, for operat-
`ing the CPU 32 as described herein. Further, the RAM 37
`stores, after being read from the hard drive, not shown, a
`table (TABLE I) mapping detected motion imparted by the
`operator of the transducer 12 into command, or control,
`signals for the workstation 14 (FIG. 1). Further it should be
`noted that the user might alter the mapping provided by the
`table using setup screen touch commands.
`[0044] Thus, the ultrasound processor 21 (FIG. 3) scan
`converts data associated with the radial scan pattern to
`generate ultrasound image data in a video format (e.g.
`Cartesian coordinate format). In one embodiment, a single
`radial scan format with possible changes in depth limits the
`number of operations for scan converting. Multiple scan
`formats and associated scan conversions may be used. Video
`filtering or processing mayalso be provided. Thus, as noted
`briefly above, the processor 21 (FIG. 3) includes the array of
`transmitting/receiving elements 18, here an array of piezo-
`electric crystals that deliver ultrasonic energy into a patient
`and receive ultrasonic echoes from the patient. Electrical
`signals representative of the echoes produced by the trans-
`ducer 12 are delivered to the beamforming network 24
`where they are selectively combined to produce an indica-
`tion ofthe echo intensity along a particulardirection or beam
`in the patient. The data produced by the beamforming
`network 24 is fed to the echo processor 26 that calculates
`echo intensity at each position along a beam and may
`calculate a Doppler shift of the echoes received along a
`particular beam. Data from the echo processor 28 is fed to
`a scan converter 28 that converts the data into a form that can
`be readily displayed on a video monitor 22.
`[0045] The data produced by the scan converter 28 is
`stored in an the RAM 37 where an additional processing,
`such as adding color, may be performedprior to displaying
`the images on a video monitor. Controlling the operation of
`the above-referenced parts are one or more central process-
`ing units, here collectively indicated by the CPU 32. The
`central processing units also recetve commands from the
`sonographer. As noted above, controls to the workstation 14
`are provided by detecting patterns of motion, to be described
`in more detail below, provided to the transducer 12 by the
`sonographer. Thus, the CPU 32 together with the image data
`stored in RAM 37 and the TABLE I stored in memory 36,
`processes the motion detection signals imparted by the
`sonographer to provide these workstation control signals.
`Recognition of the motion inputted command by the pro-
`cessor 21 results in the CPU 32 sending a signalto a light
`
`and/or buzzer 27 mounted on the workstation 14, or chang-
`ing some on-screen indicator. Activation of the light and/or
`buzzer or on screen indicator 27 provides a visual and/or
`audible indication to the sonographerthat the commandhas
`been completed.
`
`the commands or
`It should be understood that
`[0046]
`control signals provided to the workstation 14 by detecting
`patterns of motion provided to the transducer 12 by the
`sonographer may be supplemented by other tactile com-
`mands entered manually by the sonographer to the work-
`station keyboard 25 (FIG.1) or by a foot pedal 29 (FIG. 1).
`In either case, these controls allow the sonographerto adjust
`the operation of the ultrasound machine workstation 14. In
`addition, some commandor control signals maybesentafter
`some configurable delay after the pattern of motion is
`detected. This will allow controls which require a stable
`image, such as image capture, to be included in the com-
`mandtable.
`
`[0047] The transducer 12, as noted above, includes trans-
`mit and receive elements 19 (FIGS. 2A and 2B). These
`elements 19 are arranged to provide an array of elements for
`transducing between acoustical and electrical energies, such
`as a one-dimensional, 1.5D, two-dimensionalor single ele-
`ment transducer. Any of a phasedarray, linear array, curved
`array or other arrays may be used. An acoustic window, not
`shown,is disposed in the frontal portion 20 on the housing
`16 adjacent to the transducer 12.
`
`[0048] As noted above, the transducer 12 is electrically
`coupled to the workstation 14 (FIG. 1) by a cable 20. It
`should be noted that the transducer 12 might be wireless
`coupled to the workstation 14 as described in U.S. Pat. No.
`6,780,154 issued Aug. 24, 2004,
`inventors Hunt et al.,
`assigned to the same assignee as the present invention, the
`entire subject matter thereof being incorporated herein by
`reference.
`
`[0049] Referring now again to FIGS. 2A and 2B, it is
`noted that the housing 16 of the transducer 12, and more
`particularly the frontal portion 20 thereof having the array,
`here a one or two-dimensional array, of transmitting and
`receiving elements, not shown,is rectangular shaped, having
`its longer dimension along a, here Y, or azimuthal axis, and
`its shorter dimension along, here, the X, or elevation, axis,
`as indicated. An axial Z axis is thus along the length of the
`housing(i.e., an axis perpendicular to both the X and Y axes
`to provide a conventional Cartesian coordinate system for
`the transducer 12.
`
`[0050] FIG. 4A showsthe region of a scan of an image 30,
`here a sonogram, produced by placing the transducer 12 at
`one fixed position on the patient’s body, not shown.It is first
`noted that
`the transducer 12 shown in FIG. 4A has a
`one-dimensionalarray of the transmit/receive elements 19.
`Tt is next noted that the image 30 is the Y-Z plane of the
`transducer’s coordinate system described above in connec-
`tion with FIGS. 2A and 2B. Here, for this one dimensional
`array transducer we define the following directions of
`motion impartable by the sonographerto the transducer 12:
`
`[0051]
`axis;
`
`[0052]
`+Z axis;
`
`(1) a upward (U) motion is a motion along the -Z
`
`(2) a downward (D) motion is a motion along the
`
`

`

`US 2007/0078340 Al
`
`Apr. 5, 2007
`
`[0053]
`axis; and
`
`[0054]
`axis.
`
`(3) a leftward (L) motion is a motion along the -Y
`
`(4) arightward (R) motion is a motion along the +Y
`
`[0055] FIG. 4B showsthe region of a scan of an image 30
`produced by a transducer 12 having a two dimensionalarray
`of elements 19. It is noted that image 30' produced bythis
`two dimensional array transducer is a three-dimensional
`image 30'. Here, for this two-dimensional array transducer,
`wedefine the following directions of motion impartable by
`the sonographerto the transducer 12:
`
`detect patterns of changein real time images. When motion
`is detected, the timing of the motion are be usedto discrimi-
`nate between motion intendedto initiate control changes and
`motion which occurs normally during scanning. As noted
`above, patterns of direction are used to discriminate between
`motion intended to initiate control changes and motion
`which occurs normally during scanning. The combination of
`timing and direction of transducer 12 motion changes are
`used to discriminate between transducer motion intended to
`
`initiate control changes and motion which occurs normally
`during scanning.
`
`[0056]
`axis;
`
`[0057]
`axis;
`
`[0058]
`axis;
`
`[0059]
`+Z axis;
`
`[0061]
`axis.
`
`(1) a forward (F) motion is a motion along the +X
`
`(2)a backward (B) motion is a motion along the —X
`
`(3) a upward (U) motion is a motion along the -Z
`
`(4) a downward (D) motion is a motion along the
`
`(6) arightward (R) motion is a motion along the +Y
`
`[0066] More particularly, the Table I below and stored in
`memory 36 (FIG. 3) provides an exemplary repertoire of
`motions imparted to the transducer by the sonographer and
`which are interpreted by data stored in a Table II below of
`the memory 36, e.g., an EPROM, of the processor as
`command, or control signals for the workstation. Thus, the
`memory stores a table (TABLE I, below) mapping, in this
`example 14 detectable motionsof the transducer 12 each one
`of the 14 motions(1.e., identified by the designations “ID1”
`through “ID14”) corresponding to one of 14 command
`signals for the workstation 14, as indicated in FIG. 6A.It is
`(5) a leftward (L) motion is a motion along the -Y
`[0060]
`noted that each one of the exemplary patterns 33 (FIG. 6A)
`axis; and
`in TABLE|is different from merely changing the position
`of the transducer 12 to obtain a different scan view. For
`example, a sequence ofa left flick of the wrist followed by
`a right flick of the wrist is not the type of motion used to
`merely change the scan view. Further, it is noted that each
`pattern includes a sequenceofat least twoflicks of the wrist
`(each pair offlick typically occurring in a second of time or
`less). Still further, a single flick of the wrist, as shown by the
`curve 33 in FIG. 6B may be used assuming it
`is fast
`compared with the motion typically, or normally, used to
`change transducer image position shownby the curve 31 in
`FIG. 6A.
`
`The transducer is moved to the right, then back to the
`original position
`Transducer moved
`position
`to the right, back to the original
`Transducer moved
`position, then back to the right and finally back
`to the original position.
`Transducer moved
`to the left, back to the original
`position, then back to theleft, and finally back
`to the original position.
`Transducer moved
`to the right, to the left past the
`original position,
`then back right to the original
`position.
`to theleft, to ther right past the
`Transducer moved
`then back to the original position.
`original position,
`Transducer moved down, then back up.
`7 D-U
`Transducer moved up, then back down.
`8 U-D
`9 D-U-D-U Transducer moved down, then back up, then the motion
`is repeated.
`10 U-D-U-D Transducer moved up, then back down, then the motion
`is repeated.
`to the right, then down, then up,
`Transducer moved
`hen back to the original position
`Transducer moved
`to the left, then down, then up,
`hen back right to the original position
`13 R-L-D-U Transducer moved
`to the right, left, down, and back
`to the original position
`patterns of transducer motion, and hardware and/or software
`
`14 L-R-D-U_Transducer moved to theleft, right, down, and back
`up to the original position.
`to map those patterns to the activation of system controls.
`The detection of these sonographer’s imparted transducer
`motions may be performed by hardware and/or software to
`
`
`
`5 R-L-L-R
`
`6 L-R-R-L
`
`11 R-D-U-L
`
`12 L-D-U-R
`
`[0062] The ultrasound system 10 (FIG. 1) is capable of
`displaying the image 30 in either orientation, (also U/D
`inverted) it is simply an operator preference. A small sym-
`bol, not shown, is displayed on the screen 22 which corre-
`sponds to a physical notch, not shown, on the transducer 12
`housing so the operator (and anyone viewing the images
`later) can tell which way the imageis oriented. In FIGS. 4A
`and 4B the surface the patient, not shown,
`is in the X-Y
`plane, and the Z axis is “into” the patient’s body.
`
`It should be understoodthat, as is well known,the
`[0063]
`term “Linear array” refers to a one-dimensional (1D) array
`used to produce a “Linear” image, while a “Sector array” or
`“Vector array” refers to a 1D array used to produce a
`“Sector” image. The physical geometry ofthe transducersis
`similar, but vector arrays tend to be smaller. The shape of the
`image is determined by the way the systems controls the
`electrical timing of the transmit and receive signals. A third
`image format is the “Curved Linear” image, produced by a
`linear transducer with a convex curve along the azimuthal
`dimension of the transducer surface.
`
`[0064] FIG.5A showsof a motion ofthe transducer 12 by
`the sonographer along the X-axis (i.e., a forward/backward
`motion); FIG. 5B showsof a motion of the transducer 12 by
`the sonographer along the Y-axis (i.e., a right (R)/left (L)
`motion; and, FIG. 5C showsof a motion of the transducer 12
`by the sonographer along the Z-axis (1.e., an up (U)/down
`(D) motion.
`
`[0065] As noted above, the processor 21 (FIG. 3) detects
`patterns of these X, Y and/or Z sonographer imparted
`motions to provide controls to the workstation 14. The
`invention consists of software and/or hardware to detect
`
`ID Name
`
`1 R-L
`
`2 L-R
`
`3 R-L-R-L
`
`4 L-R-L-R
`
`
`
`Description
`
`
`
`
`
`TABLEI
`
`to the left, then back to the original
`
`
`
`

`

`US 2007/0078340 Al
`
`Apr. 5, 2007
`
`[0067]
`
`ID Name
`
`14 L-R-D-U
`
`
`
`
`
`be used to discriminate between motion intended to initiate
`control changes and motion which occurs normally during
`scanning. In addition, patterns of direction are be used to
`discriminate between motion intended to initiate control
`
`changes and motion which occurs normally during scanning.
`The combination of timing and direction of transducer 12
`motion changesare used to discriminate between transducer
`motion intended to initiate control changes and motion
`which occurs normally during scanning.
`
`[0070] Referring now to FIG. 7, a flow diagram of one
`methodused herein to generate workstation commandsfrom
`sonographer imparted motions to the transducer is shown.
`
`TABLEII
`
`Userinterface action
`
`
`
`Capture the image and store it in a patient database..
`1 R-L
`Start capturing image data to a movieclip.
`2 LR
`3 R-L-R-L Mouse Rightclick, bring up a menu.
`4 L-R-L-R Mouse double click, select a user interface object,
`such as a menu item.
`5 R-L-L-R
`Invoke an on-screen cursor.
`6 L-R-R-L
`Select
`the next measurement in a series of
`measurements.
`Mouseleft click
`Start automatic image gain adjustment
`9 D-U-D-U_Start VCR recording
`[0071] The scanner(i.e., scanning system) acquires image
`10 U-D-U-D Stop VCR recording
`data (Step 700). The processor 21 (FIG. 3) determines the
`Start a trace tool
`overall image motion vector (Step 702). If the determined
`Enter a calculation report screen
`overall image movement vector is greater than a predeter-
`Go to the next stage in a defined exam protocol.
`This may change a combination of imaging parameters,
`mined threshold (i.e., the motion is consistent with the flick
`stopwatch timers, image annotations, measurement
`of the sonographer’s wrist or a rapid up-down motionof the
`tools and calculation package measurements.
`transducer as distinguished from a motion consistent with
`Display the next entry in a series of pre-defined image
`the sonographer merely changing the position of the trans-
`annotation text strings.
`Enter cine review playback
`ducerto obtain a different view of the region being observed
`Start voice activation listening
`of the patient) the processor 21 acquires additional image
`Start voice annotation recording
`data and stores such data in RAM 34 (Step 704). It is noted
`Stop voice annotation recording
`that, in general, the magnitude threshold filters ou

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket