throbber

`
`
`Exhibit 1001
`
`
`Mako Surgical Corp.
`EX. 1001 Page 1
`
`Mako Surgical Corp.
`Ex. 1001
`
`Mako Surgical Corp. Ex. 1001
`
`Mako Surgical Corp. Ex. 1001
`
`

`

`(12) United States Patent
`(10) Patent N0.:
`US 6,757,582 B2
`
`Brisson et al.
`(45) Date of Patent:
`Jun. 29, 2004
`
`USOO6757582B2
`
`(54) METHODS AND SYSTEMS TO CONTROLA
`SHAPING TOOL
`
`(56)
`
`References Cited
`US. PATENT DOCUMENTS
`
`(75)
`
`Inventors: Gabriel Brisson, Pittsburgh, PA (US);
`Takeo Kanade, Pittsburgh, PA (US);
`Anthony DiGiOia, 111, Pittsburgh, pA
`(US); Branislav Jaramaz, Pittsburgh,
`PA (US)
`
`(73) Assignee: Carnegie Mellon University,
`Pittsburgh, PA (US)
`
`( * ) Notice:
`
`Subject to any disclaimer, the term of this
`patent is extended or adjusted under 35
`U.S.C. 154(b) by 0 days.
`
`(21) Appl. N05 10/427’093
`(22)
`Filed:
`Apr. 30, 2003
`
`(65)
`
`Prior Publication Data
`US 2003/0208296 A1 Nov. 6, 2003
`
`(60)
`
`(51)
`Egg
`
`Related US. Application Data
`Provisional application No. 60/377,695, filed on May 3,
`2002.
`
`7
`
`................................................ G06F 19/00
`Int. Cl.
`gfsl'dCIfs""""h"""""""" 700/186; 83973315261162:
`earc
`.................................
`,
`,
`1e
`0
`700/159, 245; 606/128; 318/568.11; 144/31;
`83/768, 367, 370; 451/5
`
`4/1987 Brumbach .................. 606/128
`4,660,573 A *
`
`9/1995 Brust et al. ........... 606/128
`5,449,363 A *
`
`. 318/568.11
`8/2000 Katoh etal.
`6,097,168 A *
`6,501,997 B1 * 12/2002 Kakmo ....................... 700/159
`6,520,228 B1 *
`2/2003 Kennedy et al.
`............. 83/768
`* cited by examiner
`
`Primary Examiner—Albert W. Paladini
`(74) Attorney, Agent, or Firm—Kevin A. Oliver; Foley
`Hoag LLP
`(57)
`
`ABSTRACT
`
`A method and system for providing control that include
`providing a workpiece that includes a target shape, provid-
`ing a cutting tool, providing a 3-D image associated with the
`workpiece, identifying the target shape within the workpiece
`image, providing a 3-D image associated with the cutting
`tool, registering the workpiece with the workpiece image,
`registering the cutting tool with the cutting tool image,
`tracking at least one of the workpiece and the cutting tool,
`transforming the tracking data based on image coordinates
`to determine a relationship between the workpiece and the
`cutting tool, and, based on the relationship, providing a
`control to the cutting tool. In one embodiment, the work-
`piece image can be represented as volume pixels (voxels)
`that can be classified and/or reclassified based on target
`~
`Shape’ waSte’ and/0r workplece'
`65 Claims, 13 Drawing Sheets
`
`CUTTING TOOL
`CUTTING
`
`SENSOR
`COLLECTION
`
`COORDINATE
`xrom
`
`CALIBRATION
`
`COLLISION/
`INTERFERENCE
`DETECTION
`
`PROBE
`
`Mako Surgical Corp. Ex. 1001
`Page 1
`
`Page 1
`
`Mako Surgical Corp. Ex. 1001
`
`

`

`US. Patent
`
`Jun. 29, 2004
`
`Sheet 1 0f 13
`
`US 6,757,582 B2
`
`GENERATE WORKPIECE IMAGE
`
`INTEGRATE TARGET SHAPE INTO WORKPIECE IMAGE
`
`“VOXILLATE” INTEGRATED WORKPIECE IMAGE
`
`GENERATE CUTTING TOOL IMAGE
`
`CALIBRATE POINT PROBE
`
`100
`
`1l2
`
`1l4
`
`1l6
`
`1 I 8
`
`1 0
`
`ASSOCIATE MARKERS WITH CUTTING TOOL AND WORKPIECE
`
`USE POINT PROBE TO REGISTER CUTTING TOOL AND WORKPIECE
`
`1 2
`
`ITERATIVELY:
`TRACK CUTTING TOOL AND WORKPIECE
`
`PROVIDE CONTROL TO CUTTING TOOL
`
`VOXILLATE INTEGRATED WORKPIECE IMAGE
`
`FIGURE 1
`
`Mako Surgical Corp. Ex. 1001
`Page 2
`
`Page 2
`
`Mako Surgical Corp. Ex. 1001
`
`

`

`US. Patent
`
`Jun. 29, 2004
`
`Sheet 2 0f 13
`
`US 6,757,582 B2
`
`114
`
`120
`
`TRACK MARKERS ON WORKPIECE AND/OR CUTTING TOOL
`
`(RECEIVE TRACKING DATA)
`
`TRANSFORM TRACKING DATA TO IMAGE COORDINATES
`
`122
`
`
`
`
`
` 124
`
` 126
`
`
`INTERSECTION DETECTION/COMPUTE CONTROL
`
`TRANSMIT CONTROL TO CUTTING TOOL AND UPDATE
`IIVIAGE/VOXELS BASED ON INTERSECTION DETECTION
`
`FIGURE 2A
`
`Mako Surgical Corp. Ex. 1001
`Page 3
`
`Page 3
`
`Mako Surgical Corp. Ex. 1001
`
`

`

`US. Patent
`
`Jun. 29, 2004
`
`Sheet 3 0f 13
`
`US 6,757,582 B2
`
`ESTABLISH GRID OF VOXELS
`
`INCORPORATE WORKPIECE IMAGE DATA TO GRID
`
`CLASSIFY VOXELS
`
`REGISTER WORKPIECE/CUTTING TOOL TO IMAGES
`
`UPDATE WORKPIECE AND/OR CUTTING TOOL IMAGES
`
`
`
`WITH TRACKING DATA
`
`
`
`
`COMPUTE CONTROL
`
`PERFORM COLLISION/INTERSECTION DETECTION/
`
`
`
`
`UPDATE IMAGE/VOXELS (CLASSIFICATION, DIVISION,
`COMBINATION, ...)/TI-U\NSMIT CONTROL
`
`130
`
`132
`
`.
`
`126A
`
`112
`
`120, 122
`
`124
`
`126
`
`FIGURE 2B
`
`Mako Surgical Corp. Ex. 1001
`Page 4
`
`Page 4
`
`Mako Surgical Corp. Ex. 1001
`
`

`

`US. Patent
`
`Jun. 29, 2004
`
`Sheet 4 0f 13
`
`US 6,757,582 B2
`
`1100
`
`m 1108 1102 1102
`lllm
`VInm-
`
`L
`
`
`
`1 10
`
`FIGURE 3A
`
`Mako Surgical Corp. Ex. 1001
`Page 5
`
`Page 5
`
`Mako Surgical Corp. Ex. 1001
`
`

`

`US. Patent
`
`Jun. 29, 2004
`
`Sheet 5 0f 13
`
`US 6,757,582 B2
`
`1100
`
`1102
`
`1102
`
` 1108
`‘22:!
`man
`
`
`
`1104
`
`
`
`1101
`
`FIGURE 3B
`
`Mako Surgical Corp. Ex. 1001
`Page 6
`
`Page 6
`
`Mako Surgical Corp. Ex. 1001
`
`

`

`
`
`
`
`
`
`
`1122 1122
`
`1122 1122
`
`1134 1134
`
`1104
`
`US. Patent
`
`Jun. 29, 2004
`
`Sheet 6 0f 13
`
`US 6,757,582 B2
`
`1120
`
`1110 1132
`
`1110 1132
`
`1132
`
`1132 1132 1132
`
`1112
`
`
`
`
`
`
`1122 1132
`
`1122 1132
`
`1132 1132 1132 1132
`
`1134 1134
`
`1134 1134 1134 1134
`
`1104
`
`1104
`
`1104
`
`FIGURE 4A
`
`Mako Surgical Corp. Ex. 1001
`Page 7
`
`Page 7
`
`Mako Surgical Corp. Ex. 1001
`
`

`

`US. Patent
`
`Jun. 29, 2004
`
`Sheet 7 0f 13
`
`US 6,757,582 B2
`
`
`
`
`
`
`
`
`1132”1132”
`
`1120
`
`1120’
`
`11029
`
`1132”1122”
`
`1132
`
`1132 1132 1132 1132 1122
`
`1122
`
`1122 1122
`
`1122 1122 1122 1122
`
`1122 1122
`
`1122 1122
`
`1122 1122 1122 1122
`
`1134 1134
`
`1134 1134
`
`1134 1134 1134 1134
`
`1104
`
`1104
`
`1104
`
`1104
`
`FIGURE 4B
`
`
`
`
`
`
`
`
`Mako Surgical Corp. Ex. 1001
`Page 8
`
`Page 8
`
`Mako Surgical Corp. Ex. 1001
`
`

`

`US. Patent
`
`Jun. 29, 2004
`
`Sheet 8 0f 13
`
`US 6,757,582 B2
`
`1132
`
`1132
`
`1122
`
`1132
`
`1122
`
`
`
`1130
`
`
`
`
`
`
`
`1132
`
`1122
`
`1132
`
`1122
`
`1122
`
`1110’1110’ 1110’1110’1132
`
`1122 1122 1122
`
`1134 1134
`
`1134 1134
`
`1134 1134 1134 1134
`
`1104
`
`1104
`
`1104
`
`1104
`
`FIGURE 4C
`
`
`
`
`
`
`
`
`Mako Surgical Corp. Ex. 1001
`Page 9
`
`Page 9
`
`Mako Surgical Corp. Ex. 1001
`
`

`

`US. Patent
`
`Jun. 29, 2004
`
`Sheet 9 0f 13
`
`US 6,757,582 B2
`
`CUTTING TOOL
`
`ELEMENT
`
`
`
`
`PROCESSOR
`
`
`
`SENSOR
`
`fl
`
`
`SENS OR
`
`COORDINATE
`
`
`
`
`
`WORKPIECE
`COLLECTION
`XFORM
`
`
`
`MODULE
`MODULE
`
`
`
`
` COLLISION/
`INTERFERENCE
`
`DETECTION
`
` PROBE
`MODULE
`
`CALIBRATION
`
`MODULE
`
`CUTTER
`
`CONTROL ‘
`
`
` IMAGE
`MODULE
`
`
`REGISTRATION
`
`MODULE
`
`
`MODULE
`
`
`
`IMAGE ACQUISITION
`
`
`
`DISPLAY
`
`‘
`
`FIGURE 5
`
`Mako Surgical Corp. Ex. 1001
`Page 10
`
`Page 10
`
`Mako Surgical Corp. Ex. 1001
`
`

`

`US. Patent
`
`Jun. 29, 2004
`
`Sheet 10 0f 13
`
`US 6,757,582 B2
`
`
`
`Mako Surgical Corp. Ex. 1001
`Page 11
`
`Page 11
`
`Mako Surgical Corp. Ex. 1001
`
`

`

`U S. Patent
`
`Jun. 29, 2004
`
`Sheet 11 0f 13
`
`US 6,757,582 B2
`
`
`
`Mako Surgical Corp. Ex. 1001
`Page 12
`
`Page 12
`
`Mako Surgical Corp. Ex. 1001
`
`

`

`US. Patent
`
`Jun. 29, 2004
`
`Sheet 12 0f 13
`
`US 6,757,582 B2
`
`
`
`Mako Surgical Corp. Ex. 1001
`Page 13
`
`Page 13
`
`Mako Surgical Corp. Ex. 1001
`
`

`

`US. Patent
`
`Jun. 29, 2004
`
`Sheet 13 0f 13
`
`US 6,757,582 B2
`
`
`
`Mako Surgical Corp. Ex. 1001
`Page 14
`
`Page 14
`
`Mako Surgical Corp. Ex. 1001
`
`

`

`US 6,757,582 B2
`
`1
`METHODS AND SYSTEMS TO CONTROLA
`SHAPING TOOL
`
`CROSS-REFERENCE TO RELATED
`APPLICATION
`
`This application claims benefit of priority to US. Provi-
`sional Patent Application Serial No. 60/377,695, filed May
`3, 2002, the contents of which are herein incorporated by
`reference in their entirety.
`
`FIELD
`
`This disclosure relates generally to controls, and more
`specifically to controlling a tool for shaping a workpiece.
`
`BACKGROUND
`
`A wide variety of applications call for objects to be
`formed from general workpieces, where the object forma-
`tion can include cutting the object from the workpiece. The
`precision required of the cut may depend on the particular
`application.
`One application that can call for high precision is the
`shaping of bone during surgery. For example, in a surgery
`such as a total knee replacement (TKR), the bones to which
`a prosthetic knee may be attached, typically the femur and
`the tibia, can be shaped to facilitate stable and effective
`implantation of the prosthesis.
`Some cutting systems achieve increased accuracy by
`fixating the workpiece, such as bone. Bone fixation can be
`accomplished using a screw(s) and/or a clamp(s) to secure
`the bone to a secure positioning brace. Fixation can be used
`for many robotic orthopedic surgery systems because such
`systems depend on a fixed target and cannot or do not track
`the target. Fixation may be used in robotic systems despite
`the risks to the patient that can include pain, infection, and
`increased recovery and rehabilitation periods caused by the
`invasive nature of fixation.
`
`Robotic and other surgical systems may be susceptible to
`failure that can occur suddenly and can cause damage or
`injury to the subject. Controllable or detectible failure may
`be detected before harm is done. Undetectable failure may
`cause damage when the system appears to be functioning
`normally. If a robot provides power drive or power assis-
`tance to an operator, failure may result when the robot
`malfunctions because the operator may be unable to react in
`time or with sufficient force to prevent the robot components
`from injuring the subject. Robotic systems may also be
`undetectable failures, since the operator may not be in full
`or even partial physical control of the robot.
`SUMMARY
`
`The disclosed methods and systems include a control
`method that includes providing a workpiece that includes a
`target shape, providing a cutting tool, providing a 3-D image
`associated with the workpiece, identifying the target shape
`within the workpiece image, providing a 3-D image asso-
`ciated with the cutting tool, registering the workpiece with
`the workpiece image, registering the cutting tool with the
`cutting tool image, tracking the workpiece and/or the cutting
`tool, transforming the tracking data based on image coor-
`dinates to determine a relationship between the workpiece
`and the cutting tool, and based on the relationship, providing
`a control to the cutting tool. The control can include an
`analog signal, a digital signal, a control to at least partially
`retract a cutting element associated with the cutting tool, a
`control to reduce the speed of a cutting element associated
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`40
`
`45
`
`50
`
`55
`
`60
`
`65
`
`2
`with the cutting tool, a control to stop a cutting element
`associated with a cutting tool, or another control.
`The methods and systems can include representing the
`workpiece image using volume pixels (voxels), and classi-
`fying the workpiece image voxels based on the target shape.
`Accordingly, based on the relationship between the cutting
`tool and the workpiece, the methods and systems can include
`re-classifying the voxels based on the relationship.
`The methods and systems can include providing an image
`based on CT scan data, X-ray data, MRI data, fiuoroscopy
`data, and/or ultrasound data. The methods and systems can
`also include classifying such image data, represented as
`three dimensional volume pixels or “voxels,” where classi-
`fying the image voxels based on the target shape includes
`distinguishing target shape voxels and workpiece voxels. In
`an embodiment, distinguishing target and workpiece voxels
`includes associating target shape voxels with the target
`shape and associating non-target shape voxels as waste.
`Color-coding voxels, such as target shape voxels associated
`with the target shape, can also be performed to distinguish
`voxels. The images and/or voxels can be displayed to a user
`to enable a user to view relative positions of the cutting tool
`and workpiece and/or target shape. In one embodiment, the
`methods and systems can include re-classifying the voxels
`based on the relationship.
`Classifying and/or re-classifying voxels can include iden-
`tifying mixture voxels that include part workpiece and part
`target shape, subdividing the mixture voxels, and iteratively
`identifying and subdividing mixture voxels to a predeter-
`mined voxel resolution. In one embodiment, mixture voxels
`can be understood to be voxels that can be associated with
`
`more than one classification, where exemplary voxel clas-
`sifications can include target, workpiece, waste, empty,
`cutting tool, cutting element, or other classifications. Sub-
`dividing the mixture voxels can be performed based on an
`octree data structure. Further, the methods and systems can
`include recombining voxels having the same classification,
`where such recombining can generally be performed based
`on neighboring voxels of the same classification.
`The methods and systems can also include a probe that
`can be calibrated and employed to register the workpiece
`and/or the cutting tool to the workpiece image and the
`cutting tool image, respectively. The disclosed tracker can
`include a tracking method and system based on providing
`one or more markers on or otherwise associated with the
`
`workpiece and/or the cutting tool. The tracker can measure
`and/or determine at least one position and at least one angle
`associated with the workpiece and/or the cutting tool, where
`in one embodiment, the tracker can track in three positions
`and three angles to provide six degrees of freedom. The
`tracked data can thus be transformed to an image coordinate
`system to allow an updating of the respective image
`positions, angles, etc.
`The image updating can also include (re)classifying vox-
`els associated with the workpiece, where the reclassification
`can be based on the tracking data associated with the
`workpiece and/or the cutting tool. Such classifying and/or
`reclassifying can include identifying voxels associated with
`the workpiece that are eliminated by the cutting tool. The
`classifying and/or reclassifying can also include identifying
`mixture voxels, subdividing the mixture voxels, and, itera-
`tively identifying and subdividing mixture voxels until
`reaching a predetermined voxel resolution. As provided
`previously, identifying mixture voxels includes identifying
`voxels having more than one classification. The subdividing
`can be based on an octree data structure. Voxel recombina-
`
`tion of voxels having the same classification can also be
`performed.
`
`Mako Surgical Corp. Ex. 1001
`Page 15
`
`Page 15
`
`Mako Surgical Corp. Ex. 1001
`
`

`

`US 6,757,582 B2
`
`3
`Accordingly, the methods and systems include providing
`a control based on determining a distance between the
`cutting tool image and the voxels classified based on the
`target shape. In one embodiment, the control can be based on
`increasing the size of the cutting tool image to determine
`whether the increased size cutting tool intersects with the
`target image. The cutting tool image can be increased by a
`fixed amount, and/or based on tracking data associated with
`the cutting tool. The control provided to the cutting tool can
`thus be based on the relationship between a cutting element
`associated with the cutting tool image, and voxels classified
`based on the target shape.
`In an embodiment, the control provided to the cutting tool
`can be based on the relationship between the cutting tool
`image and the voxels classified and/or associated with the
`target shape, where the relationship can be based on colli-
`sion detection and/or intersection detection between at least
`part of the cutting tool and voxels associated with the target
`shape.
`In one embodiment, the workpiece image can be under-
`stood to be associated with voxels that can be further
`
`associated with a three-dimensional grid of voxels, where an
`image associated with the workpiece can be incorporated
`into the grid, and grid voxels can be identified as being
`associated with the workpiece. Some of the workpiece
`voxels can thus further be associated with the target shape.
`The methods and systems include providing a control to
`the cutting tool by performing at least one of collision
`detection and intersection detection. Such control can per-
`forming at least one of collision detection and intersection
`detection between at least part of the cutting tool and the
`target shape of the workpiece image.
`In the disclosed methods and systems, identifying the
`target shape includes classifying voxels associated with the
`workpiece image as at least one of workpiece and target
`shape. Accordingly, providing control to the cutting tool can
`include performing at least one of collision detection and
`intersection detection between at least part of the cutting tool
`and the target shape voxels. Providing control can also
`include providing a control based on a threshold distance
`between the workpiece image and the cutting tool image.
`Also disclosed is a system that includes a cutting tool, a
`workpiece that includes a target shape, a tracker to provide
`tracking data associated with the cutting tool and the
`workpiece, and a controller to control the cutting tool based
`on the tracking data associated with the cutting tool and the
`tracking data associated with the workpiece. The cutting tool
`can include one or more cutting elements that can include
`one or more blade(s), one or more rotatable blade(s), one or
`more retractable blade(s), one or more water jet(s), one or
`more particulate jet(s), one or more lithotriptor(s) and/or one
`or more ultrasonic lithotriptor(s). The controller can control
`the cutting tool by providing a control to at least partially
`retract the cutting element(s), and/or at least partially reduce
`a rotation rate and/or change a cutting rate of the cutting
`element(s). The controller can transmit a control signal to
`the cutting tool, where the control signal includes an analog
`signal, a digital signal, and no signal.
`The systems can include a tracker that includes or other-
`wise provides tracking data based on at least three positions
`and at least three angles. The tracker can include one or more
`first markers associated with the workpiece, and one or more
`second markers associated with the cutting tool. The mark-
`ers or some of the markers can be one or more infrared
`
`sources, Radio Frequency (RF) sources, ultrasound sources,
`and/or transmitters. The tracker can thus be an infrared
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`40
`
`45
`
`50
`
`55
`
`60
`
`65
`
`4
`tracking system, an optical tracking system, an ultrasound
`tracking system, an inertial tracking system, a wired system,
`and/or a RF tracking system.
`The systems also include one or more images associated
`with the workpiece and at least one image associated with
`the cutting tool. The workpiece image(s) can be registered to
`the workpiece, and the cutting tool image(s) can be regis-
`tered to the cutting tool. Accordingly, the systems include a
`means to register the workpiece to the image(s) associated
`with the workpiece, and a means to register the cutting tool
`to the image(s) associated with the cutting tool. The regis-
`tration means can include a probe that can be calibrated prior
`to registration. Registration can be performed by contacting
`locations on the workpiece and/or cutting tool with the
`calibrated probe.
`The systems thus also include means to provide at least
`one image associated with the workpiece, and means to
`provide at least one image associated with the cutting tool.
`Such means can include Computer Aided Design (CAD), CT
`scan, MRI data, X-ray,
`fluoroscopy, and/or ultrasound,
`although other means can be used. The systems can update
`the images with tracking data using means to transform the
`tracking data between different coordinate systems. Such
`transformations can be mathematically effectuated.
`The systems and methods can be applied to a workpiece
`that
`includes bone, cartilage,
`tendon,
`ligament, muscle,
`connective tissue, fat, neuron, hair, skin, a tumor, and an
`organ. The cutting tool can include an endoscopic instru-
`ment.
`The controller can also include a collision detection
`module and/or an intersection detection module that can
`determine a relationship between the cutting tool and at least
`part of the workpiece.
`Disclosed is a system that includes a workpiece having a
`target shape included therein, a tracker to track at least one
`of a cutting tool and the workpiece, and, a control system,
`the control system including instructions to cause a proces-
`sor to track the cutting tool and the workpiece, to determine
`a relationship between the cutting tool and at least one of the
`workpiece and the target shape, and to provide a control to
`the cutting tool based on at least one of the relationship of
`the cutting tool and the workpiece, and the relationship of
`the cutting tool and the target shape. The control system can
`also include an image associated with the workpiece and an
`image associated with the cutting tool. The image associated
`with the workpiece can includes an image associated with
`the target shape, and/or at least part of the workpiece image
`can be designated and/or otherwise classified as being
`associated with the target shape.
`The system also includes an image registration means,
`where the image registration means registers the workpiece
`to an image associated with the workpiece, and the image
`registration means registers the cutting tool to an image
`associated with the cutting tool, and wherein the control
`system includes instructions to update at least positions of
`the workpiece image and the cutting tool image based on
`data from the tracker; and, where at
`least one of the
`relationship of the cutting tool and the workpiece, and the
`relationship of the cutting tool and the target shape, are
`based on the updated image positions.
`In the disclosed systems,
`the relationship between the
`cutting tool and the workpiece can be based on position data
`and/or angle data associated with the cutting tool(s) and/or
`the workpiece, where the position data and angle data can be
`based on the tracker. The relationship between the cutting
`tool and the target shape can thus be based on position data
`
`Mako Surgical Corp. Ex. 1001
`Page 16
`
`Page 16
`
`Mako Surgical Corp. Ex. 1001
`
`

`

`US 6,757,582 B2
`
`5
`and/or angle data associated with the cutting tool and/or the
`target shape, where the position data and angle data are
`based on the tracker. The instructions to determine a rela-
`
`tionship between the cutting tool and the target shape and/or
`workpiece can also include instructions to represent
`the
`workpiece as a group of volume pixels (voxels), classify
`voxels corresponding to the target shape, represent
`the
`cutting tool as a group of voxels, a surface model, and/or
`using constructive solid geometry or other geometric
`modeling, and, based on the tracker data, classify and/or
`update the voxels. The instructions to classify voxels cor-
`responding to the target shape can include classifying voxels
`as target shape and classifying voxels as waste, and/or
`instructions to color-code voxels corresponding to the target
`shape. In an embodiment, the workpiece can be represented
`as a surface model
`
`The disclosed methods and systems can include a control
`for a shaping tool that can be referred to herein as a cutting
`tool, and in one embodiment, is a freehand shape cutter, but
`can be understood to be a tool that can cut, shave, and/or
`grind. References herein to a shaping tool or cutting tool can
`accordingly be understood to represent a tool that can cut,
`shave, and/or grind.
`The disclosed methods and systems include a freehand
`shape cutter that includes a handheld cutting tool having a
`cutting element and a first marker. A second marker can be
`affixable to a workpiece that includes a target shape. A
`tracker can track a position of the cutting tool based on a
`position of the first marker, and also track a position of the
`workpiece based on a position of the second marker. A
`controller can control
`the cutting element based on the
`position of the cutting tool and the position of the workpiece
`to prevent
`the cutting element from invading the target
`shape.
`In one exemplary embodiment, the methods and systems
`include a method of shaping a bone by determining a target
`shape of the bone, aligning the target shape with the bone,
`providing a handheld cutting tool having a cutting element,
`tracking the bone and the cutting tool, cutting the bone with
`the cutting tool, and controlling the cutting element
`to
`prevent invasion of the cutting tool on the target shape. In
`such an embodiment, the target shape of the bone can be
`determined by creating a bone model based on geometrical
`data of the bone, and establishing the target shape based on
`the bone model.
`
`In one embodiment, the cutting tool can have six degrees
`of freedom, and the tracker can track with six degrees of
`freedom.
`
`The cutting element can include at least one of a blade, a
`rotatable blade, a retractable blade, a water jet, a particulate
`jet, a lithotriptor, and an ultrasonic lithotriptor. The control-
`ler can control the cutting element by at least one of stopping
`the cutting element, retracting the cutting element, progres-
`sively retracting the cutting element, switching off the
`cutting element, and interrupting power to the cutting ele-
`ment. The tracked and/or determined positions can be three-
`dimensional positions that can be tracked substantially
`simultaneously. Additionally and optionally,
`the positions
`can be tracked continuously.
`The target shape may be represented in the controller as
`a virtual template. The workpiece can include, for example,
`at least one of bone, cartilage, tendon, ligament, muscle,
`connective tissue, fat, neuron, hair, skin, tumor, and an organ
`that can include skin, brain, meninges, palate,
`tongue,
`esophagus, stomach, duodenum,
`jejunum,
`ileum, colon,
`liver, kidney, spleen, pancreas, ganglion, heart, artery, vein,
`
`5
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`40
`
`45
`
`50
`
`55
`
`60
`
`65
`
`6
`trachea, bronchus,
`lung,
`arteriole, venule, capillary,
`bronchiole, alveolus, blood, extremity, and a reproductive
`organ. A tumor can include a neoplasm, a benign tumor, a
`hyperplasia, a hypertropy, a dysplasia, an anaplasia, a
`metaplasia, a metastasis, and a malignant tumor.
`
`BRIEF DESCRIPTION OF THE FIGURES
`
`FIG. 1 is a block diagram of one embodiment;
`FIGS. 2A and 2B illustrate embodiments of the disclosed
`
`methods and systems;
`FIGS. 3A—3B provide schematic diagrams of embodi-
`ments of voxellation;
`FIGS. 4A—4C provide schematic diagrams of embodi-
`ments of voxellation;
`FIG. 5 is an architectural block diagram of one embodi-
`ment;
`
`FIGS. 6 and 7 depict embodiments of cutting systems;
`FIGS. 8 depicts an embodiment of a handheld cutting
`tool;
`FIGS. 9A and 9B depict an embodiment of a cutting tool
`with a retractable head;
`FIGS. 10A and 10B illustrate an exemplary cutting head;
`FIGS. 11A and 11B also illustrate an exemplary cutting
`head;
`FIG. 12 provides a cutting tool for endoscopic use; and
`FIG. 13 shows a cross-section of the FIG. 12 tool.
`
`DETAILED DESCRIPTION
`
`To provide an overall understanding, certain illustrative
`embodiments will now be described; however, it will be
`understood by one of ordinary skill in the art that the systems
`and methods described herein can be adapted and modified
`to provide systems and methods for other suitable applica-
`tions and that other additions and modifications can be made
`
`without departing from the scope of the systems and meth-
`ods described herein.
`
`Unless otherwise specified, the illustrated embodiments
`can be understood as providing exemplary features of vary-
`ing detail of certain embodiments, and therefore, unless
`otherwise specified, features, components, modules, and/or
`aspects of the illustrations can be otherwise combined,
`separated, interchanged, and/or rearranged without depart-
`ing from the disclosed systems or methods.
`The disclosed systems and methods include a methods
`and systems for controlling a cutting tool.
`In one
`embodiment, the cutting tool can be controlled relative to a
`workpiece and/or an object (target) that can be derived from
`the workpiece. Although the illustrated embodiments and
`other examples provided herein relate to surgical applica-
`tions where the workpiece can be, for example, a bone, those
`of ordinary skill in the art will recognize that the disclosed
`methods and systems relate to a system and method where
`a target shape can be developed from other workpieces,
`where the workpiece can include a material including, for
`example, wood, plastic, living tissue, ceramic, plaster, or
`other non-living materials. For surgical applications, work-
`pieces can include, for example,
`living tissue including
`bone, cadaveric grafts, or engineered tissue grafts.
`The disclosed methods and systems thus include and/or
`can be associated with a workpiece from which a target
`shape can be formed using a cutting tool to cut, grind away,
`or otherwise eliminate pieces or portions of the workpiece.
`When appropriate cuts are made and pieces or portions
`appropriately eliminated, the remaining workpiece can be
`
`Mako Surgical Corp. Ex. 1001
`Page 17
`
`Page 17
`
`Mako Surgical Corp. Ex. 1001
`
`

`

`US 6,757,582 B2
`
`7
`to the desired target shape.
`substantially similar
`Accordingly, the workpiece can be understood to include a
`target shape and waste, wherein the cutting tool can be used
`to eliminate the waste from the workpiece to leave the target
`shape. The disclosed methods and systems can thus include
`generating one or more computer models and/or a workpiece
`image that includes the target shape and a cutting tool image,
`registering the workpiece and the cutting tool to the respec-
`tive computer models and/or images, and facilitating target
`shape formation by tracking the cutting tool (and/or the
`cutting element) and the workpiece relative to the computer
`models and/or images to enable the cutting tool (and/or
`cutting element) to cut and/or eliminate those portions of the
`workpiece that are not part of the target shape. A 2D
`representation of the 3D images can be provided on a
`display.
`Ablock diagram describing the features of a method and
`system as disclosed herein can be as provided in FIG. 1. As
`previously provided, the features of FIG. 1 are not provided
`in particular order, include varying levels of detail for certain
`embodiments, and such features are presented for illustrative
`purposes. Accordingly, the illustrated features of FIG. 1 can
`be rearranged in terms of order, and as described herein,
`some features can be further detailed and/or eliminated in
`some embodiments.
`
`As FIG. 1 indicates, a workpiece image can be provided
`or otherwise generated 100 for
`input
`to a processor-
`controlled device that can include, and can be referred to
`herein, as a computer. In some embodiments, the workpiece
`image can be a three-dimensional (3-D) image and/or can be
`translated to a 3-D image for presentation. Some embodi-
`ments allow the workpiece image to be manipulated on the
`display screen by a computer user using keyboard, joystick,
`mouse, audio, and/or other commands. For example,
`in
`medical applications, the 3-D image can be provided by a
`Computer Aided Design (CAD), Computed Tomography
`(CT) Scan, Medical Resonance Imaging (MRI), and/or
`combinations of such data and/or other data that can include
`X-Ray, digital image, or other data and/or data formats. In
`an embodiment, a 3-D image can be provided by a 3-D wire
`model derived from orthogonal projections of an object, as
`described for example in Zdravkovic and Bilic, “Computer-
`assisted preoperative planning (CAPP)
`in orthopedic
`surgery,” Computer Methods and Programs in Biomedicine,
`32 (1990) 141—146. The computer can include a display and
`instructions for the processor that can cause a 2-D repre-
`sentation of the workpiece image to be presented on the
`display.
`Once the workpiece image is provided 100, a target shape
`can be integrated into the workpiece image 102. In some
`embodiments, the target shape can also be a 3-D image that
`can be integrated with the workpiece image, or the work-
`piece image may merely be modified to include designations
`of those image portions that are target shape, and those
`portions which are not target shape. For the purposes of the
`present disclosure, non-target shape areas of the workpiece
`can be referred to herein as “waste.” One method of gen-
`erating an integrated workpiece/target shape is provided in
`US. Pat. No. 6,205,411, the contents of which are herein
`incorporated by reference.
`Once the integrated workpiece/target image exists 102,
`the integrated image can be “voxellated.” The term “voxel-
`lated” as provided herein, can include a method for subdi-
`viding an image into three-dimensional (3-D) pixels, or
`“voxels,” where initially,
`the integrated image can be
`divided into voxels having resolution of, for example, 0.3
`millimeters in each of three dimensions, although such
`
`8
`resolution is merely for illustration and not limitation, and
`voxel resolution can depend on the embodiment and can
`depend, for example, on resolutions provided by a tracker
`for the disclosed methods and systems.
`In some embodiments, the voxel size and/or dimensions
`can match the cutting tool tracking system accuracy, while
`in other embodiments, the voxel size can be an integer or
`fractional multiple of the tracking system accuracy and/or
`resolution. Accordingly, workpiece image data may be resa-
`mpled to provide voxels having a desired

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket