`a2) Patent Application Publication (0) Pub. No.: US 2007/0078340 Al
`(43) Pub. Date: Apr. 5, 2007
`
`Wilcox et al.
`
`US 20070078340A1
`
`(54) METHOD AND APPARATUS FOR
`CONTROLLING ULTRASOUND IMAGING
`SYSTEMS HAVING POSITIONABLE
`TRANSDUCERS
`
`(22)
`
`Filed:
`
`Sep. 30, 2005
`
`Publication Classification
`
`(75)
`
`Inventors: Stephen D. Wilcox, Los Gatos, CA
`(US); Sankaralingam Ramraj,
`Sunnyvale, CA (US)
`
`Correspondence Address:
`STEMENS CORPORATION
`INTELLECTUAL PROPERTY DEPARTMENT
`170 WOOD AVENUE SOUTH
`
`ISELIN, NJ 08830 (US)
`
`(73) Assignee: Siemens Medical Solutions USA, Inc.
`
`(21) Appl. No.:
`
`—-11/242,161
`
`(51)
`
`Int. Cl.
`(2006.01)
`A6GIB 8/00
`(52) US. C1. eee cecssseceseessesceneneesssecsssensssenseneeneaee 600/437
`
`(57)
`
`ABSTRACT
`
`A method and system for providing an operational command
`signal to a workstation of an imaging system. The worksta-
`tion is provided imagingdata from a positionable transducer.
`The method and system convert at least one of a predeter-
`mined plurality of motion patterns imparted by an operator
`of the system to the transducer into the operational com-
`mandsignal.
`
`14
`
`
`
`Patent Application Publication Apr.5,2007 Sheet 1 of 12
`
`US 2007/0078340 Al
`
`
`
`Patent Application Publication Apr.5,2007 Sheet 2 of 12
`
`US 2007/0078340 Al
`
`15
`
`Lv.
`
`FIG. 2A
`
`20
`
`
`
`FIG. 2B
`
`
`
`US 2007/0078340 Al
`
`Patent Application Publication Apr.5,2007 Sheet 3 of 12
`
`FIG. 2C
`
`20
`
`
`
`
`CNT
`
`
`NTT
`
`TTT
`
`
`OUDDEUOREORREOOREONOOEROREROREOUELL
`OUUDEUOARAROOGROOOROERGOREGOOROUERH
`
`
`UT
`
`
`OUDEEUOARARUGROGGRREOREORROOOEOUERI
`TEEEE
`
`
`FIG. 2D
`
`
`
`Patent Application Publication Apr.5, 2007 Sheet 4 of 12
`
`US 2007/0078340 Al
`
`obeu
`
`JOSSe00i¢
`
`weeg
`
`Burwi0o4
`
`ueoS
`
`JaYaAU0D
`
`oyoz
`
`JOSSED0Jq
`
`OMION
`
`
`
`Patent Application Publication Apr.5,2007 Sheet 5 of 12
`
`US 2007/0078340 Al
`
`aD Transducer
`
`X Axis (Elevation)
`"Forward/
`
`Backward"
`
`1D Array,
`2D Image
`
`"U p"
`
`20
`
`Transducer Face
`a’
`
`"Right"
`
`Y Axis (Azimuth)
`
`"Down"
`
`FIG. 4A
`
`
`
`ZAxis(Axial)
`
`
`
`
`
`15
`
`2D Transducer
`
`12
`
`X Axis (Elevation)
`
`individual Transducer
`Elemets, 19
`
`Transducer Face
`
`
`Y Axis (Azimuth)
`
`2D Array,
`3D Image
`
`
`
`ZAxis(Axial)
`
`FIG. 4B
`
`
`
`(Axis me"?
`AyOrn Vy
`<—(f XY j )
`i /
`\
`|
`py!
`\
`\
`/
`
`/
`
`/
`/
`/
`/
`!
`/
`
`FIG. SA
`
`Patent Application Publication Apr.5,2007 Sheet 6 of 12
`
`US 2007/0078340 Al
`
`12
`
`—> “7
`
`\
`
`!
`/
`
`/
`
`/
`
`/
`
`FIG. 5C }
`
`i
`
`
`
`Patent Application Publication Apr.5,2007 Sheet 7 of 12
`
`US 2007/0078340 Al
`
`Pictograph——
`
`
`
`
`
`|
`Y
`z
`12
`
`
`
`Userinterface action
`ae
`
`
`len|<a
`F SS
`R-L-R-L
`Ys Mouse Rightclick, bring up a menu.
`
`
`
`
`WN
`AL-R-R-L
`
`
`
`
`
`
`7
`
`
`jo|@|apSS
`
`
`
`
`
`
`
`
`
`
`
`
`13)R-L-D-U
`
`GZ
`
`_R./P.
`
`
`
`
`Goto the next stage in a defined exam protocol. This may
`change a combination of imaging parameters, stopwatch
`timers, image annotations, measurementtools and calculation
`33 package measurements.
`
`Display the next entry in a series of pre-defined
`
`
`
`Explanation of pictographs:
`
`
`
`
`30_f\= Vector Image
`
`
`= Path of transducer motion
`
`
`
`
`
`
`
`
`FIG. 6A
`
`
`
`
`
`
`
`
`of measurements.
`it Mouseleft click.
`
`SS
`
`9S
`
`S
`
`iW
`NJ
`a=
`
`Toyo[=EPIERIE
`
`R-L-L-R
`
`U-D
`
`
`
`
`
`
`
`
`
`
`
`Patent Application Publication Apr.5,2007 Sheet 8 of 12
`
`US 2007/0078340 Al
`
`Y Axis Position
`
`Triggerlevels resetif pattern
`not detected within
`timout period
`
`Sonographer
`command
`signal motion
`
`-— ee
`
`
`
`Motion that
`occurs during
`normal scanning
`
`
`
`
`
`Timer Timout
`
`Time
`
`Timerstarted at
`first level crossing
`
`FIG. 6B
`
`
`
`Patent Application Publication Apr.5, 2007 Sheet 9 of 12
`
`US 2007/0078340 Al
`
`Scanneracquires image data
`
`Determine overall image movement vector
`
`
`
`
`Yes
`
`
`Motion
`
`greater than
`threshold?
`
`
`Acquire additional image data
`
`
`Determine additional overall
`image motion vector
`
`threshold?
`
`
`
`
`Matches
`
`a knownpattern
`
`template?
`
`
`
`
`
` Motion
`greater than
`
`Yes
`
`Compare motion vector pattern
`to pattern templates
`
`
`LookupUserInterface Action for pattern
`
`Yes
`
`Send commandto userinterface software
`
`700
`
`702
`
`704
`
`706
`
`708
`
`710
`
`712
`
`FIG. 7
`
`
`
`Patent Application Publication Apr.5,2007 Sheet 10 of 12
`
`US 2007/0078340 Al
`
`
`Yes
`
`
`
`
`
`
` Motion
`
`
`greater than
`threshold?
`
`
`
`
`Timer
`
`
`not expired?
`
`Yes
`
`Acquire additional image data
`
`Determine additional overall image motion vector
`
` Motion
`
`
`
`greater than
`hreshold?
`
`
`Yes
`
`To Fig. 8B
`
`FIG. 8A
`
`810
`
`}—812
`
`
`
`Patent Application Publication Apr.5,2007 Sheet 11 of 12
`
`US 2007/0078340 Al
`
`FIG. 8
`
`FIG. 8A
`
`FIG. 8B
`
`From Fig. 8A
`
`
`
` Timer
`not expired?
`
`Yes
`
`Compare motion vectorpattern to
`pattern templates
`
`
`
`
`Matches
`a knownpattern
`template?
`
`Yes
`
`
`
`814
`
`816
`
`818
`
`
`
`
`
`
`Lookup UserInterface Action for pattern
`
`
`Send command to userinterface software
`
`FIG. 8B
`
`
`
`Patent Application Publication Apr.5, 2007 Sheet 12 of 12
`
`US 2007/0078340 Al
`
`image
`
`Sub-image
`segments
`
`Sub-image
`LZ.
`segments used
`detect overall
`image motion
`
`Sub-
`
`image
`
`a&&@7)
`
`oOwosBw
`
`detect overall
`image motion
`
`
`
`
`US 2007/0078340 Al
`
`Apr. 5, 2007
`
`METHOD AND APPARATUS FOR CONTROLLING
`ULTRASOUND IMAGING SYSTEMS HAVING
`POSITIONABLE TRANSDUCERS
`
`
`
`TECHNICAL FIELD
`
`[0001] This invention relates generally to methods and
`apparatus for controlling imaging systems and more par-
`ticularly for controlling imaging systems having position-
`able transducers.
`
`BACKGROUND
`
`Asis knowninthe art, one type of imaging system
`[0002]
`is an ultrasound imaging systems. A conventional ultrasound
`imaging system includesa positional transducer, typically a
`sonographer handheld transducer, coupled to a large pro-
`cessing and display workstation or operator interface. The
`frontal portion of the transducer includes an array ofultra-
`sonic elements which transmit and receive ultrasonic energy
`for imaging a selected region of a patient. The received
`ultrasonic energy is converted to electric signals by the
`transducer and passed to the workstation. The workstation
`detects, filters and otherwise processes the information to
`generate a two- or three-dimensional representation of the
`scanned region.
`
`[0003] The sonographer supplies the control signals for
`the workstation. Such control signals are typically supplied
`by the sonographer’s free, or non-transducer carrying hand.
`Scanning situations in both the examination room and other
`locations often require the sonographer to be in awkward
`positions for simultaneously reaching the controls with the
`free hand and placing the frontal portion of the transducer in
`the proper position on the patient’s body. One technique
`suggested to solve this problem is through voice activation;
`however, such technique may be proneto error and requires
`a speech recognition learning phase for each sonographer.
`Another technique suggested to provide the central signals
`to the workstation is through a sonographer actuated foot
`pedal; however such is not practical for all scanning situa-
`tions.
`
`SUMMARY
`
`In accordance with the present invention, a method
`[0004]
`is provided for providing an operational commandsignals,
`sometimes herein referred to as control signals, to a work-
`station of an imaging system. The workstation is provided
`imaging data from a positionable transducer. One method
`includes converting at least one of a plurality of predeter-
`mined motion patterns imparted by an operatorof the system
`to the transducer into the operational commandsignals.
`
`[0005] With such method, the operator is provided with a
`wayto control the workstation without taking a handoff of
`the transducer, or relying on voice control or foot actuated
`controls. The method reduces the number of times the
`operator must touch controls on the workstation.
`
`[0006] Another method includes parting the transducer
`head into multiple regions (in one embodiment, less than
`four regions) and interpreting the reception of the signals
`from such regions into operational commandsignals.
`
`the converting comprises
`In one embodiment,
`[0008]
`detecting at least one of the predetermined motion patterns
`and converting such detected motion patterns into a corre-
`sponding one of the operational commandsignals.
`
`In one embodiment, the detecting comprises com-
`[0009]
`paring a sequence of images formed by the system.
`
`Inone embodiment, the methodincludes determin-
`[0010]
`ing from the sequence of images whether the motion
`imparted to the transducer waseither a repositioning of the
`transducer to producea different image to be observed by the
`operator or a motion imparted to produce the corresponding
`one of the commandsignals to the workstation.
`
`In one embodiment, such determining includes
`[0011]
`comparing types of motions imparted bythe operation.
`
`In one embodiment, such determining includes
`[0012]
`comparing imparted motion with a level threshold.
`
`In one embodiment, such determining includes
`[0013]
`comparing imparted motion with a time duration threshold.
`
`In one embodiment, a method is provided for
`[0014]
`providing control signals to a workstation of an imaging
`system, such workstation being provided imaging data from
`a positionable transducer. The method includes detecting
`patterns of motion of the transducer, and converting the
`patterns to the control signals.
`
`In one embodiment, the detection is performed by
`[0015]
`detecting patterns of changein real time images provided by
`the system.
`
`In one embodiment, timing of the motion is used to
`[0016]
`discriminate between motion intendedto provide the control
`signals and motion normally occurring during scanning.
`
`In one embodiment, patterns of direction of the
`[0017]
`transducer motion are used to discriminate between motion
`
`intended to provide the control signals and motion normally
`occurring, during scanning.
`
`In one embodiment, a combination of patterns of
`[0018]
`direction of the transducer motion and timing of the motion
`are used to discriminate between motion intendedto provide
`the control signals and motion normally occurring during
`scanning.
`
`In one embodiment, an imaging system is provided
`[0019]
`having a workstation and a positionable transducer for
`providing imaging data to the workstation. The workstation
`responds to the operational commandsignals. The worksta-
`tion includes a memory for storing a table mapping detected
`motion of the transducer into the commandsignals.
`
`the workstation includes a
`In one embodiment,
`[0020]
`processor programmedto detect at least one of a predeter-
`minedplurality of motion patterns and convert such detected
`one of the motion patterns into the operational command
`signals.
`
`In one embodiment, the transducer has deposed
`[0021]
`within a housing thereof motion sensors.
`
`In one embodiment, sensors disposed remote from
`[0022]
`the transducer sense motion of the transducer.
`
`[0007] A third method includes converting detections of
`predetermined echo signatures into operational command
`signals.
`
`In one embodiment, an imaging system is provided
`[0023]
`having a workstation and a positionable transducer for
`providing imaging data to the workstation. The workstation
`
`
`
`US 2007/0078340 Al
`
`Apr. 5, 2007
`
`responds to control signals. The workstation includes a
`processor for detecting patterns of motion of the transducer
`and converting the patterns to the control signals.
`
`[0024] The details of one or more embodiments of the
`invention are set forth in the accompanying drawings and
`the description below. Other features, objects, and advan-
`tages of the invention will be apparent from the description
`and drawings, and from the claims.
`
`DESCRIPTION OF DRAWINGS
`
`FIG.1 is a sketch of an imaging system according
`[0025]
`to the invention;
`
`[0026] FIG. 2A is an isometric sketch of a transducer used
`in the system of FIG. 1;
`
`[0027] FIG. 2B is a front elevation view of the frontal
`portion of the transducer of FIG. 2A;
`
`[0028] FIG. 2C is an isometric sketch of a transducer
`adapted for use in the system of FIG. 1 according to one
`embodimentof the invention;
`
`[0029] FIG. 2D is a front elevation view of the frontal
`portion of a 2D array transducer;
`
`[0030] FIG. 3 is a block diagram of a processor used to
`process imaging data from the transducer of FIGS. 2A and
`2B to generate an image for an operator of a workstation
`used in the system of FIG. 1 and to process such data to
`generate control signal for operation of the workstation;
`
`[0031] FIG. 4A is a diagram showing a coordinate system
`for a one-dimensionalarray transducer used in the system of
`FIG. 1, such coordinate system indicating and defining
`operator motion of the transducer;
`
`FIG.4B is a diagram showing a coordinate system
`[0032]
`for a two-dimensional array transducer used in the system of
`FIG. 1, such coordinate system indicating and defining
`operator motion of the transducer;
`
`[0033] FIGS. 5A, 5B and 5C show various patterns of
`motion which may be imparted to the transducer by the
`operator and then image data which is processed by the
`processor of FIG. 3 to generate the control signally for the
`workstation of the system of FIG. 1;
`
`[0034] FIG. 6A is a table showing the relationship
`between a repertoire of motions impartable by the operator
`to the transducer and the workstation control
`signals
`intended by the operator resulting from such motions;
`
`[0035] FIG. 6B shows a comparison between normal
`transducer motion occurring during scanning and motion
`used to initiate a commandsignalto the workstation of FIG.
`1;
`
`[0036] FIG. 7 is a flow diagram of a process used by the
`processor of FIG. 3 in generating the workstation control
`signals from images generated by the transducer;
`
`[0037] FIGS. 8, 8A and 8B are flow diagrams in more
`detail of the process of FIG. 7 used by the processor of FIG.
`3 in generating the workstation control signals from images
`generated by the transducer;
`
`[0038] FIG. 9 is a diagram showing an image obtained by
`the processor of FIG. 3 divided into sub-image segments by
`the processor and used to detect overall image motion; and
`
`[0039] FIG. 10 is a diagram showing an imageobtained by
`the processor of FIG. 3 divided into sub-image segments by
`the processor with near-field segments used to detect overall
`image motion.
`
`[0040] Like reference symbols in the various drawings
`indicate like elements.
`
`DETAILED DESCRIPTION
`
`[0041] FIG. 1 shows an imaging system 10, here an
`ultrasound imaging system for medical diagnostics. The
`system 10 includes a positionable, here handheld,
`image
`processing ultrasound device, here a transducer 12 shown in
`more detail in FIGS. 2A and 2B, and a multi-use display
`device, or operator interface, herein sometime collectively
`referred to as a workstation 14. The handheld transducer 12
`obtains ultrasound data and formats the ultrasound data for
`transmission to the workstation 14, here via a cable 15.
`Controls to the workstation 14 are provided by detecting
`patterns of motion, to be described in more detail below,
`provided, or imparted, to the transducer 12 by an operator of
`the system 10, typically by a sonographer. Suffice it to say
`here that the pattern of motions may include: an up-down or
`down-up(i.e., axial) motion of the sonographer’s transducer
`hand holding wrist (i.e., an up flick of the wrist followed by
`a downflick of the wrist or a sequence of down-up flicks of
`the wrist); a left-rightor right-left (1.e., azimuthal) motion of
`the sonographer’s transducer hand holding wrist (i.e., left-
`rightflicks or right-left flicks of the wrist); for 2D arrays, a
`forward-and-backward (1.e., elevational) motion of the
`sonographer’s transducer hand holding wrist; a inward-
`outward motion towards and away from the patient’s body,
`or visa versa, or any combination thereof. The detection of
`these sonographer’s imparted transducer motions may be
`performed by hardware and/or software to detect patterns of
`change in real time images. When motion is detected, the
`timing of the motion is used to discriminate between motion
`intended to initiate control changes and motion which occurs
`normally during scanning. In addition, patterns of direction
`are be used to discriminate between motion intended to
`initiate control changes and motion which occurs normally
`during scanning. The combination of timing anddirection of
`transducer 12 motion changes are used to discriminate
`between transducer motion intended to initiate control
`
`changes and motion which occurs normally during scanning.
`
`[0042] The transducer 12 includes a housing 16 (FIG. 2A)
`adapted to be easily handheld, such as for example, being
`less than 8 inches in any dimension and/or having an
`ergonomic shape for holding in a operator’s hand. The
`housing 16 comprises plastic, rubber, metal, other materials
`now knownorlater developed, or combinations thereof. In
`one embodiment shownin FIG. 2A,the housing 16 is shaped
`for ergonomic use or holding by the operator(e.g., sonog-
`rapher) by having a generally round or curved circumference
`handle serving as a grip for the sonographer’s hand.
`
`[0043] The handheld transducer 12 includes conventional
`ultrasoundcircuitry, not shown, within the housing 16. Thus,
`the ultrasound circuitry includes, in the frontal portion 20
`thereof (FIGS. 2A and 2B) an array of ultrasonic elements
`19 which transmit and receive ultrasonic energy for imaging
`a patient, not shown. It
`is noted that FIG. 2B is for a
`one-dimensionalarray transducer and FIG. 2D is for a 2D
`array transducer. Here, the transducer’s transmit and receive
`
`
`
`US 2007/0078340 Al
`
`Apr. 5, 2007
`
`elements 19 in the frontal portion 20 are arranged in an
`elongated array along a the longeraxis, here the Y axis of the
`rectangular shaped patient interfacing surface, 1.e., the fron-
`tal portion 20, of the housing 16. The elements 20 in the
`frontal portion 20 are, as shown in FIG.3 fed to a display 22
`of the workstation 14 (FIG. 1) serially through: a beam-
`forming network 24, an echo processor 26, a scan converter
`28, and an image processor 31 in a conventional manner.
`The beamforming network 24, echo processor 26, scan
`converter 28, image processor 31 and display 22 are con-
`trolled by a central processing unit (CPU) 32 coupled to a
`random access memory RAM37. The CPU 32 operates in
`accordance with program instructions stored in a ROM 34,
`or in RAM 37,or in flash memory not shown, or on a hard
`drive device, not shown. A memory 36, here an erasable, or
`other type of programmable semiconductor memory, here a
`read only memory (ROM)is provided for storing a com-
`puter, here microprocessor, executable program, for operat-
`ing the CPU 32 as described herein. Further, the RAM 37
`stores, after being read from the hard drive, not shown, a
`table (TABLE I) mapping detected motion imparted by the
`operator of the transducer 12 into command, or control,
`signals for the workstation 14 (FIG. 1). Further it should be
`noted that the user might alter the mapping provided by the
`table using setup screen touch commands.
`[0044] Thus, the ultrasound processor 21 (FIG. 3) scan
`converts data associated with the radial scan pattern to
`generate ultrasound image data in a video format (e.g.
`Cartesian coordinate format). In one embodiment, a single
`radial scan format with possible changes in depth limits the
`number of operations for scan converting. Multiple scan
`formats and associated scan conversions may be used. Video
`filtering or processing mayalso be provided. Thus, as noted
`briefly above, the processor 21 (FIG. 3) includes the array of
`transmitting/receiving elements 18, here an array of piezo-
`electric crystals that deliver ultrasonic energy into a patient
`and receive ultrasonic echoes from the patient. Electrical
`signals representative of the echoes produced by the trans-
`ducer 12 are delivered to the beamforming network 24
`where they are selectively combined to produce an indica-
`tion ofthe echo intensity along a particulardirection or beam
`in the patient. The data produced by the beamforming
`network 24 is fed to the echo processor 26 that calculates
`echo intensity at each position along a beam and may
`calculate a Doppler shift of the echoes received along a
`particular beam. Data from the echo processor 28 is fed to
`a scan converter 28 that converts the data into a form that can
`be readily displayed on a video monitor 22.
`[0045] The data produced by the scan converter 28 is
`stored in an the RAM 37 where an additional processing,
`such as adding color, may be performedprior to displaying
`the images on a video monitor. Controlling the operation of
`the above-referenced parts are one or more central process-
`ing units, here collectively indicated by the CPU 32. The
`central processing units also recetve commands from the
`sonographer. As noted above, controls to the workstation 14
`are provided by detecting patterns of motion, to be described
`in more detail below, provided to the transducer 12 by the
`sonographer. Thus, the CPU 32 together with the image data
`stored in RAM 37 and the TABLE I stored in memory 36,
`processes the motion detection signals imparted by the
`sonographer to provide these workstation control signals.
`Recognition of the motion inputted command by the pro-
`cessor 21 results in the CPU 32 sending a signalto a light
`
`and/or buzzer 27 mounted on the workstation 14, or chang-
`ing some on-screen indicator. Activation of the light and/or
`buzzer or on screen indicator 27 provides a visual and/or
`audible indication to the sonographerthat the commandhas
`been completed.
`
`the commands or
`It should be understood that
`[0046]
`control signals provided to the workstation 14 by detecting
`patterns of motion provided to the transducer 12 by the
`sonographer may be supplemented by other tactile com-
`mands entered manually by the sonographer to the work-
`station keyboard 25 (FIG.1) or by a foot pedal 29 (FIG. 1).
`In either case, these controls allow the sonographerto adjust
`the operation of the ultrasound machine workstation 14. In
`addition, some commandor control signals maybesentafter
`some configurable delay after the pattern of motion is
`detected. This will allow controls which require a stable
`image, such as image capture, to be included in the com-
`mandtable.
`
`[0047] The transducer 12, as noted above, includes trans-
`mit and receive elements 19 (FIGS. 2A and 2B). These
`elements 19 are arranged to provide an array of elements for
`transducing between acoustical and electrical energies, such
`as a one-dimensional, 1.5D, two-dimensionalor single ele-
`ment transducer. Any of a phasedarray, linear array, curved
`array or other arrays may be used. An acoustic window, not
`shown,is disposed in the frontal portion 20 on the housing
`16 adjacent to the transducer 12.
`
`[0048] As noted above, the transducer 12 is electrically
`coupled to the workstation 14 (FIG. 1) by a cable 20. It
`should be noted that the transducer 12 might be wireless
`coupled to the workstation 14 as described in U.S. Pat. No.
`6,780,154 issued Aug. 24, 2004,
`inventors Hunt et al.,
`assigned to the same assignee as the present invention, the
`entire subject matter thereof being incorporated herein by
`reference.
`
`[0049] Referring now again to FIGS. 2A and 2B, it is
`noted that the housing 16 of the transducer 12, and more
`particularly the frontal portion 20 thereof having the array,
`here a one or two-dimensional array, of transmitting and
`receiving elements, not shown,is rectangular shaped, having
`its longer dimension along a, here Y, or azimuthal axis, and
`its shorter dimension along, here, the X, or elevation, axis,
`as indicated. An axial Z axis is thus along the length of the
`housing(i.e., an axis perpendicular to both the X and Y axes
`to provide a conventional Cartesian coordinate system for
`the transducer 12.
`
`[0050] FIG. 4A showsthe region of a scan of an image 30,
`here a sonogram, produced by placing the transducer 12 at
`one fixed position on the patient’s body, not shown.It is first
`noted that
`the transducer 12 shown in FIG. 4A has a
`one-dimensionalarray of the transmit/receive elements 19.
`Tt is next noted that the image 30 is the Y-Z plane of the
`transducer’s coordinate system described above in connec-
`tion with FIGS. 2A and 2B. Here, for this one dimensional
`array transducer we define the following directions of
`motion impartable by the sonographerto the transducer 12:
`
`[0051]
`axis;
`
`[0052]
`+Z axis;
`
`(1) a upward (U) motion is a motion along the -Z
`
`(2) a downward (D) motion is a motion along the
`
`
`
`US 2007/0078340 Al
`
`Apr. 5, 2007
`
`[0053]
`axis; and
`
`[0054]
`axis.
`
`(3) a leftward (L) motion is a motion along the -Y
`
`(4) arightward (R) motion is a motion along the +Y
`
`[0055] FIG. 4B showsthe region of a scan of an image 30
`produced by a transducer 12 having a two dimensionalarray
`of elements 19. It is noted that image 30' produced bythis
`two dimensional array transducer is a three-dimensional
`image 30'. Here, for this two-dimensional array transducer,
`wedefine the following directions of motion impartable by
`the sonographerto the transducer 12:
`
`detect patterns of changein real time images. When motion
`is detected, the timing of the motion are be usedto discrimi-
`nate between motion intendedto initiate control changes and
`motion which occurs normally during scanning. As noted
`above, patterns of direction are used to discriminate between
`motion intended to initiate control changes and motion
`which occurs normally during scanning. The combination of
`timing and direction of transducer 12 motion changes are
`used to discriminate between transducer motion intended to
`
`initiate control changes and motion which occurs normally
`during scanning.
`
`[0056]
`axis;
`
`[0057]
`axis;
`
`[0058]
`axis;
`
`[0059]
`+Z axis;
`
`[0061]
`axis.
`
`(1) a forward (F) motion is a motion along the +X
`
`(2)a backward (B) motion is a motion along the —X
`
`(3) a upward (U) motion is a motion along the -Z
`
`(4) a downward (D) motion is a motion along the
`
`(6) arightward (R) motion is a motion along the +Y
`
`[0066] More particularly, the Table I below and stored in
`memory 36 (FIG. 3) provides an exemplary repertoire of
`motions imparted to the transducer by the sonographer and
`which are interpreted by data stored in a Table II below of
`the memory 36, e.g., an EPROM, of the processor as
`command, or control signals for the workstation. Thus, the
`memory stores a table (TABLE I, below) mapping, in this
`example 14 detectable motionsof the transducer 12 each one
`of the 14 motions(1.e., identified by the designations “ID1”
`through “ID14”) corresponding to one of 14 command
`signals for the workstation 14, as indicated in FIG. 6A.It is
`(5) a leftward (L) motion is a motion along the -Y
`[0060]
`noted that each one of the exemplary patterns 33 (FIG. 6A)
`axis; and
`in TABLE|is different from merely changing the position
`of the transducer 12 to obtain a different scan view. For
`example, a sequence ofa left flick of the wrist followed by
`a right flick of the wrist is not the type of motion used to
`merely change the scan view. Further, it is noted that each
`pattern includes a sequenceofat least twoflicks of the wrist
`(each pair offlick typically occurring in a second of time or
`less). Still further, a single flick of the wrist, as shown by the
`curve 33 in FIG. 6B may be used assuming it
`is fast
`compared with the motion typically, or normally, used to
`change transducer image position shownby the curve 31 in
`FIG. 6A.
`
`The transducer is moved to the right, then back to the
`original position
`Transducer moved
`position
`to the right, back to the original
`Transducer moved
`position, then back to the right and finally back
`to the original position.
`Transducer moved
`to the left, back to the original
`position, then back to theleft, and finally back
`to the original position.
`Transducer moved
`to the right, to the left past the
`original position,
`then back right to the original
`position.
`to theleft, to ther right past the
`Transducer moved
`then back to the original position.
`original position,
`Transducer moved down, then back up.
`7 D-U
`Transducer moved up, then back down.
`8 U-D
`9 D-U-D-U Transducer moved down, then back up, then the motion
`is repeated.
`10 U-D-U-D Transducer moved up, then back down, then the motion
`is repeated.
`to the right, then down, then up,
`Transducer moved
`hen back to the original position
`Transducer moved
`to the left, then down, then up,
`hen back right to the original position
`13 R-L-D-U Transducer moved
`to the right, left, down, and back
`to the original position
`patterns of transducer motion, and hardware and/or software
`
`14 L-R-D-U_Transducer moved to theleft, right, down, and back
`up to the original position.
`to map those patterns to the activation of system controls.
`The detection of these sonographer’s imparted transducer
`motions may be performed by hardware and/or software to
`
`
`
`5 R-L-L-R
`
`6 L-R-R-L
`
`11 R-D-U-L
`
`12 L-D-U-R
`
`[0062] The ultrasound system 10 (FIG. 1) is capable of
`displaying the image 30 in either orientation, (also U/D
`inverted) it is simply an operator preference. A small sym-
`bol, not shown, is displayed on the screen 22 which corre-
`sponds to a physical notch, not shown, on the transducer 12
`housing so the operator (and anyone viewing the images
`later) can tell which way the imageis oriented. In FIGS. 4A
`and 4B the surface the patient, not shown,
`is in the X-Y
`plane, and the Z axis is “into” the patient’s body.
`
`It should be understoodthat, as is well known,the
`[0063]
`term “Linear array” refers to a one-dimensional (1D) array
`used to produce a “Linear” image, while a “Sector array” or
`“Vector array” refers to a 1D array used to produce a
`“Sector” image. The physical geometry ofthe transducersis
`similar, but vector arrays tend to be smaller. The shape of the
`image is determined by the way the systems controls the
`electrical timing of the transmit and receive signals. A third
`image format is the “Curved Linear” image, produced by a
`linear transducer with a convex curve along the azimuthal
`dimension of the transducer surface.
`
`[0064] FIG.5A showsof a motion ofthe transducer 12 by
`the sonographer along the X-axis (i.e., a forward/backward
`motion); FIG. 5B showsof a motion of the transducer 12 by
`the sonographer along the Y-axis (i.e., a right (R)/left (L)
`motion; and, FIG. 5C showsof a motion of the transducer 12
`by the sonographer along the Z-axis (1.e., an up (U)/down
`(D) motion.
`
`[0065] As noted above, the processor 21 (FIG. 3) detects
`patterns of these X, Y and/or Z sonographer imparted
`motions to provide controls to the workstation 14. The
`invention consists of software and/or hardware to detect
`
`ID Name
`
`1 R-L
`
`2 L-R
`
`3 R-L-R-L
`
`4 L-R-L-R
`
`
`
`Description
`
`
`
`
`
`TABLEI
`
`to the left, then back to the original
`
`
`
`
`
`US 2007/0078340 Al
`
`Apr. 5, 2007
`
`[0067]
`
`ID Name
`
`14 L-R-D-U
`
`
`
`
`
`be used to discriminate between motion intended to initiate
`control changes and motion which occurs normally during
`scanning. In addition, patterns of direction are be used to
`discriminate between motion intended to initiate control
`
`changes and motion which occurs normally during scanning.
`The combination of timing and direction of transducer 12
`motion changesare used to discriminate between transducer
`motion intended to initiate control changes and motion
`which occurs normally during scanning.
`
`[0070] Referring now to FIG. 7, a flow diagram of one
`methodused herein to generate workstation commandsfrom
`sonographer imparted motions to the transducer is shown.
`
`TABLEII
`
`Userinterface action
`
`
`
`Capture the image and store it in a patient database..
`1 R-L
`Start capturing image data to a movieclip.
`2 LR
`3 R-L-R-L Mouse Rightclick, bring up a menu.
`4 L-R-L-R Mouse double click, select a user interface object,
`such as a menu item.
`5 R-L-L-R
`Invoke an on-screen cursor.
`6 L-R-R-L
`Select
`the next measurement in a series of
`measurements.
`Mouseleft click
`Start automatic image gain adjustment
`9 D-U-D-U_Start VCR recording
`[0071] The scanner(i.e., scanning system) acquires image
`10 U-D-U-D Stop VCR recording
`data (Step 700). The processor 21 (FIG. 3) determines the
`Start a trace tool
`overall image motion vector (Step 702). If the determined
`Enter a calculation report screen
`overall image movement vector is greater than a predeter-
`Go to the next stage in a defined exam protocol.
`This may change a combination of imaging parameters,
`mined threshold (i.e., the motion is consistent with the flick
`stopwatch timers, image annotations, measurement
`of the sonographer’s wrist or a rapid up-down motionof the
`tools and calculation package measurements.
`transducer as distinguished from a motion consistent with
`Display the next entry in a series of pre-defined image
`the sonographer merely changing the position of the trans-
`annotation text strings.
`Enter cine review playback
`ducerto obtain a different view of the region being observed
`Start voice activation listening
`of the patient) the processor 21 acquires additional image
`Start voice annotation recording
`data and stores such data in RAM 34 (Step 704). It is noted
`Stop voice annotation recording
`that, in general, the magnitude threshold filters ou