WO2004069040A2 - Method and apparatus for computer assistance with intramedullary nail procedure - Google Patents

Method and apparatus for computer assistance with intramedullary nail procedure Download PDF

Info

Publication number
WO2004069040A2
WO2004069040A2 PCT/US2004/003068 US2004003068W WO2004069040A2 WO 2004069040 A2 WO2004069040 A2 WO 2004069040A2 US 2004003068 W US2004003068 W US 2004003068W WO 2004069040 A2 WO2004069040 A2 WO 2004069040A2
Authority
WO
WIPO (PCT)
Prior art keywords
surgeon
nail
images
femur
leg
Prior art date
Application number
PCT/US2004/003068
Other languages
French (fr)
Other versions
WO2004069040A3 (en
Original Assignee
Z-Kat, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Z-Kat, Inc. filed Critical Z-Kat, Inc.
Publication of WO2004069040A2 publication Critical patent/WO2004069040A2/en
Publication of WO2004069040A3 publication Critical patent/WO2004069040A3/en

Links

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/10Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges for stereotaxic surgery, e.g. frame-based stereotaxis
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B17/00Surgical instruments, devices or methods, e.g. tourniquets
    • A61B2017/00017Electrical control of surgical instruments
    • A61B2017/00207Electrical control of surgical instruments with hand gesture control or hand gesture recognition
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • A61B90/37Surgical systems with images on a monitor during operation
    • A61B2090/376Surgical systems with images on a monitor during operation using X-rays, e.g. fluoroscopy

Definitions

  • the present invention relates generally to computer-assisted surgery systems and surgical navigation systems.
  • Image-based surgical navigation systems display the positions of surgical tools with respect to preoperative (prior to surgery) or intraoperative (during surgery) image data sets.
  • Two and three dimensional image data sets are used, as well as time-variant images data (i.e. multiple data sets taken at different times).
  • Types of data sets that are primarily used include two-dimensional fluoroscopic images and three-dimensional data sets include magnetic resonance imaging (MBS) scans, computer tomography (CT) scans, positron emission tomography (PET) scans, and angiographic data.
  • Intraoperative images are typically fluoroscopic, as a C-arm fluoroscope is relatively easily positioned with respect to patients and does not require that a patient be moved. Other types of imaging modalities require extensive patient movement and thus are typically used only for preoperative and postoperative imaging.
  • the most popular navigation systems make use of a tracking or localizing system to track tools, instruments and patients during surgery. These systems locate in predefined coordinate space specially recognizable markers that are attached or affixed to, or possibly inherently a part of, an object such as an instrument or a patient. Markers can take several forms, including those that can be located using optical (or visual), electromagnetic, radio or acoustic methods. Furthermore, at least in the case of optical or visual systems, location of an object's position may be based on intrinsic features or landmarks that, in effect, function as recognizable markers. Markers will have a known, geometrical arrangement with respect to, typically, an end point and/or axis of the instrument.
  • objects can be recognized at least in part from the geometry of the markers (assuming that the geometry is unique), and the orientation of the axis and location of endpoint within a frame of reference deduced from the positions of the markers.
  • Present-day tracking systems are typically optical, functioning primarily in the infrared range. They usually include a stationary stereo camera pair that is focused around the area of interest and sensitive to infrared radiation. Markers emit infrared radiation, either actively or passively. An example of an active marker is light-emitting diodes (LEDs). An example of a passive marker is a reflective marker, such as ball-shaped marker with a surface that reflects incident infrared radiation. Passive systems require a an infrared radiation source to illuminate the area of focus.
  • a magnetic system may have a stationary field generator that emits a magnetic field that is sensed by small coils integrated into the tracked tools.
  • CAS systems are capable of continuously tracking, in effect, the position of tools (sometimes also called instruments).
  • tools sometimes also called instruments.
  • a system is able to continually superimpose a representation of the tool on the image in the same relationship to the anatomy in the image as the relationship of the actual tool to the patient's anatomy.
  • the coordinate system of the image data set must be registered to the relevant anatomy of the actual patient portions of the of the patient's anatomy in the coordinate system of the tracking system. There are several known registration methods.
  • CAS systems that are capable of using two-dimensional image data sets
  • multiple images are usually taken from different angles and registered to each other so that a representation of the tool or other object (which can be real or virtual) can be, in effect, projected into each image.
  • a representation of the tool or other object which can be real or virtual
  • its projection into each image is simultaneously updated.
  • the images are acquired with what is called a registration phantom in the field of view of the image device, hi the case of a two dimensional fluoroscopic images, the phantom is a radio-translucent body holding radio- opaque fiducials having a known geometric relationship.
  • Knowing the actual position of the fiducials in three dimensional space when each of the images are taken permits determination of a relationship between the position of the fiducials and their respective shadows in each of the images. This relationship can then be used to create a transform for mapping between points in three-dimensional space and each of the images.
  • the relative positions of tracked tools with respect to the patient's anatomy can be accurately indicated in each of the images, presuming the patient does not move after the image is acquired, or that the relevant are portions of the patient's anatomy is are tracked.
  • the invention is generally directed to improved computer-implemented methods and apparatus for further reducing the invasiveness of surgical procedures, eliminating or reducing the need for external fixtures in certain surgical procedures, and/or improving the precision and/or consistency of surgical procedures.
  • the invention finds particular advantage in orthopedic procedures involving implantation of devices, though it may also be used in connection with other types of surgical procedures.
  • a surgeon encounters or has to overcome several problems during insertion of an intramedullary nail ('TM nail”), an elongated rod-shaped prosthetic device, into the canal of a fractured femur.
  • 'TM nail intramedullary nail
  • problems include matching the leg length of the injured leg with the well leg of the patient, improper rotation of the injured leg, and unpredictable flexing of the distal end of the nail.
  • fluoroscopic images are taken frequently during the procedure, thus exposing the patient and operating room personnel to radiation.
  • implantation of the LM nail using traditional methods requires use of an extra pin for determining the version of the leg for proper alignment of the rod, as well as use of a special, radio-translucent drill so that fluoroscopic images can be captured during insertion of screws into the distal end of the femur to secure the distal end of the nail.
  • FIG. 1 is a block diagram of an exemplary computer-assisted surgery system
  • FIG. 2 is a simple diagram of a patient having a fractured femur and prepared for surgery
  • FIG. 3 is a flow chart of basic steps of an application program for assisting with or guiding the planning and execution of a surgical procedure and navigation during the procedure;
  • FIG. 4 is a flow chart of basic set-up steps for an application for assisting with planning of, and navigation during, an intramedullary nail procedure;
  • FIG. 5 is a flow chart of basic steps of a reference determination portion of the planning phase of the application of FIG. 4;
  • FIGS. 6 A is a more detailed flow chart of basic steps of reference dimensions and a nail determination portion of a phase of the application of FIG. 4;
  • FIGS. 6B is a more detailed flow chart of planning injured leg for determination of fracture site, length and anteversion for the application of FIG. 4;
  • FIG. 7 is a detailed flow chart of a navigation/execution phase of the application of FIG. 4.
  • FIGS. 8-27 are representative screens of graphical user interface pages displayed by the computer-assisted surgery system of FIG. 1 during use of the application of FIG. 4.
  • surgeon include any user of a computer-assisted surgical system, a surgeon being typically a primary user.
  • FIG. 1 is a block diagram of an exemplary computer-assisted surgery (CAS) system
  • Computer-assisted surgery system (CAS) 10 comprises a display device 12, an input device 14, and a processor-based system 16, for example, a computer.
  • Display device 12 may be any display device now known or later developed for displaying two-dimensional and/or three-dimensional diagnostic images, for example, a monitor, a touch screen, a wearable display, a projection display, a head-mounted display, stereoscopic views, a holographic display, a display device capable of displaying image(s) projected from an image projecting device, for example, a projector, and/or the like.
  • Input device 14 may be any input device now known or later developed, for example, a keyboard, a mouse, a trackball, a trackable probe and/or the like.
  • the processor-based system is preferably programmable and includes one or more processors 16a, working memory 16b for temporary program and data storage that will be used primarily by the processor, and storage for programs and data, preferably persistent, such as a disk drive.
  • Removable media storage device 18 can also be used to store programs and/or transfer to or from the transfer programs.
  • Tracking system 22 continuously determines, or tracks, the position of one or more trackable markers disposed on, incorporated into, or inherently a part of surgical tools or instruments 20 with respect to a three-dimensional coordinate frame of reference.
  • CAS system 10 is programmed to be able to determine the three-dimensional coordinates of an endpoint or tip of a tool and, optionally, its primary axis using predefined or known (e.g. from calibration) geometrical relationships between trackable markers on the tool and the end point and/or axis of the tool.
  • a patient, or portions of the patient's anatomy can also be tracked by attachment of arrays of trackable markers.
  • the CAS system can be used for both planning surgical procedures (including planning during surgery) and for navigation. It is therefore preferably programmed with software for providing basic image-guided surgery functions, including those necessary determining the position of the tip and axis of instruments and for registering a patient and preoperative and/or intraoperative diagnostic image data sets to the coordinate system of the tracking system.
  • the programmed instructions for these functions are indicated as core CAS utilities 24. These capabilities allow the relationship of a tracked instrument to a patient to be displayed and constantly updated in real time by the CAS system overlaying a representation of the tracked instrument on one or more graphical images of the patient's internal anatomy on display device 12.
  • the graphical images are constructed from one or more stored image data sets 26 acquired from diagnostic imaging device 28.
  • Imaging device may be a fluoroscope, such as a C-arm fluoroscope, capable of being positioned around a patient lying on an operating table. It may also be a MR, CT or other type of imaging device in the room or permanently located elsewhere. Where more than one image is shown, as when multiple fluoroscopic images are simultaneously displayed of display device 12, the representation of the tracked instrument or tool is coordinated between the different images.
  • CAS system can be used in some procedures without the diagnostic image data sets, with only the patient being registered. Thus, the CAS system need not support the use of diagnostic images in some applications - i.e. an imageless application.
  • the CAS system may be used to run application- specific programs 30 that are directed to assisting a surgeon with planning and/or navigation during specific types of procedures.
  • the application programs may display predefined pages or images corresponding to specific steps or stages of a surgical procedure.
  • a surgeon may be automatically prompted to perform certain tasks or to define or enter specific data that will permit, for example, the program to determine and display appropriate placement and alignment of instrumentation or implants or provide feedback to the surgeon.
  • Other pages may be set up to display diagnostic images for navigation and to provide certain data that is calculated by the system for feedback to the surgeon.
  • the CAS system could also communicate information in ways, including using audibly (e.g.
  • a CAS system may feedback to a surgeon information whether he is nearing some object or is on course with a audible sound or by application of a force or other tactile sensation to the surgeon's hand.
  • the program may automatically detect the stage of the procedure by recognizing/identifying the instrument picked up by a surgeon and move immediately to the part of the program in which that tool is used.
  • Application data 32 - data generated or used by the application may also be stored processor-based system.
  • Various types of user input methods can be used to improve ease of use of the CAS system during surgery.
  • One example is the use of speech recognition to permit a doctor to speak a command.
  • Another example is the use of a tracked object to sense a gesture by a surgeon, which is interpreted as an input to the CAS system. The meaning of the gesture could further depend on the state of the CAS system or the current step in an application process executing on the CAS system.
  • a gesture may instruct the CAS system to capture the current position of the object.
  • One way of detecting a gesture is to occlude temporarily one or more of the trackable markers on the tracked object (e.g. a probe) for a period of time, causing loss of the CAS system's ability to track the object.
  • a temporary visual occlusion of a certain length (or within a certain range of time), coupled with the tracked object being in the same position before the occlusion and after the occlusion, would be interpreted as an input gesture.
  • a visual or audible indicator that a gesture has been recognized could be used to provide feedback to the surgeon.
  • tracking system 22 in combination with one or more trackable data input devices 34.
  • defined input areas which can be two-dimensional or three-dimensional. These defined input areas are visually indicated on the trackable input device so that a surgeon can see them.
  • the input areas may be visually defined on an object by representations of buttons, numbers, letters, words, slides and/or other conventional input devices.
  • the geometric relationship between each defined input area and the trackable input device is known and stored in processor-based system 16.
  • the processor can determine when another trackable object touches or is in close proximity a defined input area and recognize it as an indication of a user input to the processor-based systems.
  • the processor-based system will recognize the tool near the defined input area and treat it as a user input associated with that defined input area.
  • representations on the trackable user input correspond with user input selections (e.g. buttons) on a graphical user interface on display device 12.
  • the trackable input device may be formed on the surface of any type of trackable device, including devices used for other purposes.
  • representations of user input functions for graphical user interface are visually defined on a rear, flat surface of a base of a tool calibrator.
  • Processor-based system 16 is, in one example, a programmable computer that is programmed to execute only when single-use or multiple-use software is loaded from, for example, removable media.
  • the software would include, for example the application program 30 for use with a specific type of procedure. Media storing the application program can be sold bundled with disposable instruments specifically intended for the procedure.
  • the application program would be loaded into the processor-based system and stored there for use during one (or a defined number) of procedures before being disabled.
  • the application program need not be distributed with the CAS system.
  • application programs can be designed to work with specific tools and implants and distributed with those tools and implants.
  • the most current core CAS utilities may also be stored with the application program.
  • FIG. 2 is intended to be a representative patient with a representative fractured femur.
  • the representative patient 200 represented by a head 202, torso 204, arm 206, leg 208 and knee 210.
  • a femur 212 Indicated by dashed lines in the upper leg, above the knee, is a femur 212 that is fractured and separated into two pieces, which will be referred to as the proximal fragment 214 and distal fragment 216 to correspond with the proximal end 218 and distal end 220 of the femur.
  • a trackable marker array 212 which can be tracked by the CAS system 10 (FIG.
  • the CAS system assists a surgeon in performing an BVI nail implantation by executing a process 300 that has three basic phases: set-up phase 302, planning phase 304 and navigation phase 306.
  • the set-up phase involves the surgeon specifying to the process which type of BVI nail to be used which leg is to be operated on, type of fracture, instruments and/or tools to be tracked during the procedure, and model fluoroscope to be used, which leg is to be operated on, type of fracture, instruments, and/or tools during the process and certain other options such as determining to image and plan using the uninjured leg.
  • the set-up phase allows for skipping certain steps during the navigation or execution stage so that it flows more efficiently to the surgeon's preferences or needs.
  • the planning phase involves using fluoroscopic images to gather reference information on leg version (rotation angle) and length from the surgeon and to select nail dimensions, and placement and length of screws used to secure the nail.
  • the navigation or execution stage tracks the surgeon's instruments and trackable markers implanted in or attached to the patient's femur and provides alignment information and feedback on version and length.
  • Process 300 or parts thereof, preferably display a series of pages corresponding to stages or sub-procedures, each page being set up to display directions and information (including images) relevant to the stage of the procedure.
  • the CAS system may in addition to the pages or in place of the pages, communicate some or all of this information by other means, including audible and haptic means.
  • the process may constrain what a surgeon does in terms of the ordering of certain steps, the process preferably follows the surgeon, rather than requiring the surgeon to follow the process.
  • process 300 some steps maybe performed out of sequence or repeated.
  • the surgeon may indicate to the process the stage he or she is in or wants to go to. This may be done through user input or by the process automatically recognizing when the surgeon has either finished a stage or is preparing to go to another stage (not necessarily the next stage) by, for example, the surgeon picking up an instrument used in a particular stage and showing it to the cameras of the tracking system. Details of the process 300 will be described with reference to representative examples of screens from such pages, shown in FIGS. 8-27. These screens contemplate use of IM nails for a specific vendor. However, the process and concepts embodied or represented by the pages are not limited to any specific vendor, and aspects thereof may be employed in connection with surgical planning and guidance systems for similar types of implants.
  • step 402 asks the surgeon to identify or select which of a plurality of IM nail types or families will be used. This information is used for representing the LM nail on images taken of the patient's leg and providing feedback to the surgeon on the position of the nail during the nail insertion. If the process is set up for only one type of nail, this step may be skipped.
  • FIG. 8 is a screen of a representative example of such a page, in this case showing four families of IM nails from a particular vendor.
  • the process requests the surgeon to specify which leg is injured, what type of fluoroscope will be used, and whether the uninjured leg of the patient will be used in planning.
  • FIG. 9 is an example of a graphical interface displaying the options for selection by the surgeon.
  • a fluoroscopic images has certain advantages, other types of images can be used in place of, or in addition to, the fluoroscopic images, including without limitation preoperative three-dimensional data sets such as CT and MRI scans.
  • the surgeon is asked to specify application-specific tools that he will use during the procedure that can be or will be tracked. Surgeons may prefer to use different tools for a given step, and this step permits the surgeon to select the tool of choice so that the CAS system can properly track it.
  • the application may display a different page at a given step, or display pages in a different order, based on selection of the tool.
  • a surgeon may, for example, elect not to use a tool during a given step, or not have it tracked. The process will adjust as necessary to accommodate the preferences to avoid forcing a surgeon to find ways to bypass steps or alter presentation of the pages.
  • the CAS system is typically programmed or set up to operate with a probe and other basic tools that a surgeon may use.
  • the surgeon is given a list of the tool or tools that the application can track, from which he may select.
  • FIG. 10 shows an example of a page that displays the tools that the application is capable of or set up to track for the basic steps of the surgical procedure. The display permits the surgeon to visually select the tool (or not to have a tool) and verify the selection.
  • the tool listing in the illustrated example is also grouped by basic stages of the process. In the example, options are given for the tool that will be used for defining an entry point in the femur prior to insertion of the nail, the tool used for nail insertion, the instrument used for drilling holes to insert screws for locking the distal end of the nail, and other tools that the surgeon may want to use.
  • the program will not expect to receive an indication for the entry point and will not attempt to display the selected point on a diagnostic image. If, for example, a surgeon selects a power drill instead of a hand drill for distal locking, the CAS system will automatically assume that it is tracking a power drill during the distal locking step.
  • the CAS system calibrates the selected fluoroscope using known methods.
  • the interface for this step is illustrated in FIG. 11.
  • Steps 414, 416, 418 and 420 direct the acquisition of certain fluoroscopic images during the procedure, followed by registration of those images using known methods. If the surgeon specified that the well leg would be used for reference, images of the well leg are acquired in addition to images of the injured leg. Exemplary screen shots of the pages corresponding to the acquisition and registration of the well leg and injured leg are shown in FIGS. 12 and 13, respectively. What images are needed or desirable are listed and identified with respect to the list when they are acquired. In the illustrated examples, the required or desirable images are listed with reference to target areas 1202 defined on diagrams 1204 of a femur. The diagrams show a femur from an anterior/posterior and from a medial/lateral view.
  • A/P and M/L fluoroscopic image pair are around the midshaft of the femur.
  • Area 1202 may consist of two fracture sites, depending upon the setup phase. This allows for handling a compound fracture and allows the surgeon to image the proximal and distal fractures in separate shots.
  • Each A/P and M/L fluoroscopic image pair is preferably shown in two, side-by-side windows on the display. During image acquisition, window 1206 displays a current image from the fluoroscope.
  • the planning stage starts with a process 500.
  • certain reference information namely a reference length and version, are determined at step 502, based on information indicated on the images by the surgeon.
  • the well leg may also be used to at least initially determine appropriate nail length and diameter at step 504. Using the well leg to determine this information may be desirable in the event it is difficult to determine this information from the injured leg.
  • screw length placement and length may also be determined.
  • FIGS. 6A and 6B illustrate in the planning stage in greater detail.
  • Steps 602 to 608 involve determination of a reference length and version of the leg or femur.
  • the surgeon is prompted to indicate in the acquired images of the well leg certain anatomical landmarks, preferably the center of the femoral head, the axis of the femoral neck and shaft, an axis that extends transverse to the condyles at the posterior-most points of the condyles (the trans-epicondylar axis).
  • anatomical landmarks preferably the center of the femoral head, the axis of the femoral neck and shaft, an axis that extends transverse to the condyles at the posterior-most points of the condyles (the trans-epicondylar axis).
  • other recognizable landmarks could be used for calculating a reference length and/or version.
  • a "bull's-eye" marker 1402 is superimposed on A/P image 1404 and M/L image 1406 for assisting the surgeon in identifying the center of the generally spherical femoral head in both images.
  • This bull's-eye marker is a two-dimensional projection of a series of nested, virtual spheres in the three dimensional space of the patient.
  • the process moves automatically to step 604.
  • the process displays A/P and lateral images of the distal end of the femur.
  • the surgeon indicates on the images a marker for services as a reference point for determining a reference length for the femur.
  • the program stores this information.
  • the reference length is calculated using the references marked on the proximal end of the femur and the reference marked on the distal end of the femur.
  • the program also prompts, using, for example, directions displayed on the displayed page, and receives from the surgeon at step 406 the position and orientation for the trans-epicondylar axis of the femur.
  • FIG. 15 is an example of a screen used in these steps.
  • the acquired A P and lateral images, 1502 and 1504, respectively, of the distal end of the well femur are displayed, along with a reference line 1506.
  • the surgeon manipulates the position and orientation of the reference line in A/P image 1502 so that it is aligned with the trans-epicondylar axis.
  • the surgeon manipulates the reference line 1506 so that it is positioned on the posterior most points of the condyles.
  • a reference version is calculated at step 608.
  • the trans-epicondylar axis also serves as a reference point for calculating a reference length.
  • true A/P and lateral views of the distal end of the femur should be acquired, or they should be least taken at the substantially same angles as the A P and lateral images of the distal end of the injured femur.
  • the reference length and version are stored and displayed in area 1508 of the screen.
  • the version may be calculated using a true lateral image of the distal end and placing a reference point on the knee center in both the A/P and lateral images.
  • Steps 608, 610, 611, 612, 614 and 616 assist the surgeon with selecting a nail of appropriate length and screw dimensions using the well leg.
  • the surgeon indicates, with respect to A/P and M/L images of the distal and proximal ends of the femur, end points for the nail.
  • the process automatically determines the distance between the end points and then it selects and displays on the images a representation of the closest standard length nail.
  • steps 610 and 611 screw placement and dimensions for the proximal end of the nail and the placement of the nail end are indicated with respect to the uninjured leg.
  • a representation of the closest standard nail to the indications is then displayed at step 612.
  • FIG. 16 is a representative screen from an example of a user interface page displayed on the CAS system implementing these steps.
  • the page includes the stored A/P image 1602 and M/L image 1604 of the proximal end of the well-leg femur.
  • Superimposed on this image is a marker, in the form of a crosshair graphic 1606, for marking the estimated proximal end of the nail that will be implanted in the other (injured) leg.
  • a representation 1608 of the proximal end of the nail is also preferably superimposed on the two images, along with representations 1610 of screws that will be inserted through the proximal end of the femur and nail once the nail is fully inserted.
  • the surgeon is permitted to change, shift, rotate and move the representation of the screws in order to check its fit.
  • the page includes inputs for changing the position of the cross-hairs and representations.
  • FIG. 17 is a representative screen of a page for the surgeon to mark an estimated location for the tip of the nail.
  • the page includes the stored A/P image 1602 and M/L image 1606 of the distal end of the well- leg femur.
  • the program then provides an estimated nail length and displays the two closest standard lengths for the type of nail being used on line 1710. The surgeon selects the desired length and the program moves the nail representation to the correct length for the surgeon to confirm that selected length.
  • the diameter of the nail is estimated using at step 618, the midshaft (isthmus) of the well or injured leg.
  • FIG. 18 assumes that well-leg images were acquired at the midshaft of that leg's femur.
  • the diameter of the canal of the femur, through which the nail will be inserted, is its narrowest at the isthmus.
  • the page instructs the surgeon to place a reference marker 1806 along the canal of the femur and then select the best matching diameter from a list.
  • the width of the reference marker which is a projection of a virtual, cylindrical object (corresponding generally to a diameter of a nail) in the three-dimensional patient space into the two- dimensional fluoroscopic images, changes.
  • step 620 the surgeon is prompted to mark in the images showing the edge of the fracture at the canal of the femur.
  • a representative screen of the page displayed for this step is shown in FIG. 19.
  • the stored A/P image 1902 and the stored lateral M/L image 1904 of the midshaft of the injured femur is displayed, and cross-hairs marker 1906 is also displayed and can be moved by the surgeon to mark the edge of the fracture.
  • a representation 1908 of the nail is also superimposed for the surgeon to check and, if necessary, change the estimated nail diameter that will fit through the canal at the point of fracture.
  • FIG. 20 is a representative screen of the page displayed for this step. Its display includes the stored A/P and M/L images of the proximal end of the femur in windows 2002 and 2004. Like other pages, it includes written instructions prompting the surgeon to mark certain landmarks, namely, the femoral head and neck using a bull's eye marker 2006 and 2008, just as in FIG. 14.
  • step 606 the same landmarks used in marking the distal end of the well femur in step 606 are marked at step 624 by the surgeon and stored for use in calculating reference length and version for comparison to the well leg.
  • a representative screen of a display page for this step is shown in FIG. 21. It includes the stored A/P and M/L images 2102 and 2104 for the distal end of the injured leg. It also includes an A/P shot of the distal end of the well femur 2106 for reference to ensure proper marking of the landmarks on the images 2102 and 2104.
  • the reference line is shown on image 2106 in the position marked by the surgeon at step 606.
  • trans-epicondylar axis is marked on the images with reference line 2108 and stored.
  • steps 628 and 630 where the surgeon indicates to the process the entry point for the nail and desired position of the nail head and the screws that lock the nail head.
  • FIG. 22 a representative screen of an example of a page for receiving this information from the surgeon, the stored A/P and M/L images 2202 and 2204 of the distal end of the injured femur are displayed and overlaid with a representation 2206 of the previously selected nail and the locking screws 2208.
  • the nail head 2210 which defines the entry point for the nail into the femur, is also indicated.
  • the surgeon shifts and rotates the representation of the nail so that it fits properly in the canal and the locking screws extend up the neck of the femur shown in the images.
  • the representations of the screws are fixed to the representation of the nail, and rotate and shift with it. When the surgeon is satisfied with the placement of the nail and locking screws, this information is stored.
  • FIG. 23 A representative screen of an exemplary page that may be displayed at this step is shown in FIG. 23.
  • a list 2302 of selected tools is displayed.
  • a surgeon selects each tool on the list for calibration.
  • the tip and, optionally, axis of each tool is calculated with respect to a known point on a calibration fixture according to known methods.
  • the calibration information is stored by the CAS system so that the relationship between the displayed representation of the tool and the diagnostic images is the same as the relationship between the actual tool and the patient.
  • steps 702, 704, 706 and 708 involve guiding the surgeon to the correct entry point for inserting the nail.
  • FIG. 24 the previously acquired and stored A P and M/L images 2402 and 2404 are displayed.
  • the entry point is also marked with markers 2406.
  • the point of the tool selected/or use in forming the entry in this case an awl, is continuously tracked by the CAS system and a representation of the position of the tip of the tool displayed on the images.
  • the CAS system is continuous tracking relative changes in positions of the two fragments using trackable marker array 224 (see FIG. 2) attached to each fragment. The arrays are attached prior to registration of the images with the patient so that registration is not lost due to movement of either femoral fragment.
  • the CAS system will compensate for movement of the fragments when displaying the position of tracked tools on the diagnostic images on the CAS system's display.
  • illustrations of the objects being tracked are displayed, in this example, and awl and a trackable marker array. These illustrations also provide an indication to the surgeon if the tool or the marker array are out of the field of view of the camera by displaying, in the illustrated embodiment, a red outline on respective images in area 2414. This function is also present in subsequent tracking steps.
  • window or area 2407 the relative positions and orientations of the proximal and femoral fragments of the fractured femur are indicated by representations 2410 and 2408.
  • This window is preferably displayed during steps 708 and 714.
  • Displayed in area 2412 is reference length and version infonnation that is continuously calculated based on the relative positions of the fragments. This tracking is possible due to the known relationship between trackable marker array and the reference landmarks specified on the fragment. At the time when the landmarks on each fragment were specified, the positions of the trackable markers were also stored, thereby permitting the relative relationship to be determined.
  • the referenced lengths and version are calculated based on the relative positions of the two trackable markers.
  • the instrument used for reaming could be tracked to and its position could also be displayed on the images to ensure that it successfully bridges the fracture and enters the canal of the other fragment. Since the reaming device must be flexible and is located inside the femur, optical tracking cannot be used. Magnetic tracking, though less precise, could be employed.
  • a tool for inserting the nail referred to as a nail inserter.
  • the inserter is tracked by the CAS system. Bringing the nail inserter into view of the tracking system signals the application process to move to the next step, namely, to step 710.
  • the geometric relationship between the tool and the nail is known from the calibration step performed earlier. Therefore, by tracking the tool inserter, which remains outside the patient, the position of the nail is known.
  • FIG. 25 the stored A/P and M/L images of the proximal end of the femur are displayed. Also displayed on images using representation 2506 is the current position of the nail and the screws as the nail is being inserted and rotated. The nail insertion tool is tracked. The position of the nail and screws is determined from the position of the nail insertion tool and the geometric relationship between the nail insertion tool and nail.
  • window 2407 displays the representations of the two fragments 2408 and 2410, of the femur in the relative positions and calculated reference lengths and versions 2412. The surgeon will use the nail and screw representations to ensure that the screws are correctly aligned with the femoral neck.
  • the representations of the locking screws can be used as guides for drilling and inserting the screws.
  • FIG. 26 is an example of a page for guiding the surgeon in capturing the images.
  • the current image for the fluoroscope is shown in window 2602. If the image is acceptable, it is stored and shown in window 2604.
  • the shots or images to be acquired are, in this example, graphically illustrated in area 2606.
  • the second set of stored A/P and M/L images of the distal end of the femur should clearly show the screw holes in the distal end of the nail.
  • the lateral image needs to be a true lateral image relative to the nail.
  • the CAS system preferably automatically displays a screen or page similar to the one of FIG. 27 and performs steps 714 and 716.
  • the page of FIG. 27 includes the stored A/P image 2702 and lateral image 2704 of the distal end of the nail.
  • a representation 2706 of the instrument being used for the insertion is superimposed on the images.
  • a representation 2708 of the locking screw on the end of the instrument is also superimposed.
  • the surgeon is prompted to specify whether to archive data generated by the procedure for later reference.
  • the CAS system archives the data as directed, such as to a disk drive or removable media. This step is not illustrated.
  • Embodiments of the present invention may be implemented in software, hardware, application logic or a combination of software, hardware and application logic.
  • the software, application logic and/or hardware may reside on processor-based system 16 or on a removable storage medium. If desired, part of the software, application logic and/or hardware may reside on processor-based system 16 and part of the software, application logic and/or hardware may reside on the removable storage medium.

Abstract

A specially-programmed, computer-assisted surgery system (10) is used to reduce the number of fluoroscopic images required to be taken during the course of a intramedullary nail procedure (300), eliminates the need for a Steinman pin, and assists the surgeon in properly aligning and securing the nail during insertion.

Description

METHOD AND APPARATUS FOR COMPUTER ASSISTANCE WITH ΓNTRAMEDULLARY NAIL PROCEDURE
TECHNICAL FIELD OF THE INVENTION
The present invention relates generally to computer-assisted surgery systems and surgical navigation systems.
BACKGROUND OF THE INVENTION
Image-based surgical navigation systems display the positions of surgical tools with respect to preoperative (prior to surgery) or intraoperative (during surgery) image data sets. Two and three dimensional image data sets are used, as well as time-variant images data (i.e. multiple data sets taken at different times). Types of data sets that are primarily used include two-dimensional fluoroscopic images and three-dimensional data sets include magnetic resonance imaging (MBS) scans, computer tomography (CT) scans, positron emission tomography (PET) scans, and angiographic data. Intraoperative images are typically fluoroscopic, as a C-arm fluoroscope is relatively easily positioned with respect to patients and does not require that a patient be moved. Other types of imaging modalities require extensive patient movement and thus are typically used only for preoperative and postoperative imaging.
The most popular navigation systems make use of a tracking or localizing system to track tools, instruments and patients during surgery. These systems locate in predefined coordinate space specially recognizable markers that are attached or affixed to, or possibly inherently a part of, an object such as an instrument or a patient. Markers can take several forms, including those that can be located using optical (or visual), electromagnetic, radio or acoustic methods. Furthermore, at least in the case of optical or visual systems, location of an object's position may be based on intrinsic features or landmarks that, in effect, function as recognizable markers. Markers will have a known, geometrical arrangement with respect to, typically, an end point and/or axis of the instrument. Thus, objects can be recognized at least in part from the geometry of the markers (assuming that the geometry is unique), and the orientation of the axis and location of endpoint within a frame of reference deduced from the positions of the markers. Present-day tracking systems are typically optical, functioning primarily in the infrared range. They usually include a stationary stereo camera pair that is focused around the area of interest and sensitive to infrared radiation. Markers emit infrared radiation, either actively or passively. An example of an active marker is light-emitting diodes (LEDs). An example of a passive marker is a reflective marker, such as ball-shaped marker with a surface that reflects incident infrared radiation. Passive systems require a an infrared radiation source to illuminate the area of focus. A magnetic system may have a stationary field generator that emits a magnetic field that is sensed by small coils integrated into the tracked tools.
Most CAS systems are capable of continuously tracking, in effect, the position of tools (sometimes also called instruments). With knowledge of the position of the relationship between the tool and the patient and the patient and image data sets, a system is able to continually superimpose a representation of the tool on the image in the same relationship to the anatomy in the image as the relationship of the actual tool to the patient's anatomy. To obtain these relationships, the coordinate system of the image data set must be registered to the relevant anatomy of the actual patient portions of the of the patient's anatomy in the coordinate system of the tracking system. There are several known registration methods.
In CAS systems that are capable of using two-dimensional image data sets, multiple images are usually taken from different angles and registered to each other so that a representation of the tool or other object (which can be real or virtual) can be, in effect, projected into each image. As the position of the object changes in three dimensional space, its projection into each image is simultaneously updated. In order to register two or more two-dimensional data images together, the images are acquired with what is called a registration phantom in the field of view of the image device, hi the case of a two dimensional fluoroscopic images, the phantom is a radio-translucent body holding radio- opaque fiducials having a known geometric relationship. Knowing the actual position of the fiducials in three dimensional space when each of the images are taken permits determination of a relationship between the position of the fiducials and their respective shadows in each of the images. This relationship can then be used to create a transform for mapping between points in three-dimensional space and each of the images. By knowing the positions of the fiducials with respect to the tracking system's frame of reference, the relative positions of tracked tools with respect to the patient's anatomy can be accurately indicated in each of the images, presuming the patient does not move after the image is acquired, or that the relevant are portions of the patient's anatomy is are tracked. A more detailed explanation of registration of fluoroscopic images and coordination of representations of objects in patient space superimposed in the images is found in United States Patent 6,198,794 of Peshkin, et al., entitled "Apparatus and method for planning a stereotactic surgical procedure using coordinated fluoroscopy".
SUMMARY OF THE INVENTION
The invention is generally directed to improved computer-implemented methods and apparatus for further reducing the invasiveness of surgical procedures, eliminating or reducing the need for external fixtures in certain surgical procedures, and/or improving the precision and/or consistency of surgical procedures. The invention finds particular advantage in orthopedic procedures involving implantation of devices, though it may also be used in connection with other types of surgical procedures.
For example, a surgeon encounters or has to overcome several problems during insertion of an intramedullary nail ('TM nail"), an elongated rod-shaped prosthetic device, into the canal of a fractured femur. These problems include matching the leg length of the injured leg with the well leg of the patient, improper rotation of the injured leg, and unpredictable flexing of the distal end of the nail. To reduce the incidence of malrotation of the leg, fluoroscopic images are taken frequently during the procedure, thus exposing the patient and operating room personnel to radiation. Furthermore, implantation of the LM nail using traditional methods requires use of an extra pin for determining the version of the leg for proper alignment of the rod, as well as use of a special, radio-translucent drill so that fluoroscopic images can be captured during insertion of screws into the distal end of the femur to secure the distal end of the nail.
To address one or more of these problems, various aspects of a specially- programmed, computer-assisted surgery system are used to reduce the number of fluoroscopic images required to be taken, especially during the course of the procedure, eliminate the need for a Steinman pin, and assist the surgeon in properly aligning and securing the nail during insertion. A preferred embodiment of such an application for programming a computer-assisted surgery system is described below. BRIEF DESCRIPTION OF THE DRAWINGS
For a more complete understanding of the present invention, the objects and advantages thereof, reference is now made to the following descriptions taken in connection with the accompanying drawings in which: FIG. 1 is a block diagram of an exemplary computer-assisted surgery system;
FIG. 2 is a simple diagram of a patient having a fractured femur and prepared for surgery;
FIG. 3 is a flow chart of basic steps of an application program for assisting with or guiding the planning and execution of a surgical procedure and navigation during the procedure;
FIG. 4 is a flow chart of basic set-up steps for an application for assisting with planning of, and navigation during, an intramedullary nail procedure;
FIG. 5 is a flow chart of basic steps of a reference determination portion of the planning phase of the application of FIG. 4; FIGS. 6 A is a more detailed flow chart of basic steps of reference dimensions and a nail determination portion of a phase of the application of FIG. 4;
FIGS. 6B is a more detailed flow chart of planning injured leg for determination of fracture site, length and anteversion for the application of FIG. 4;
FIG. 7 is a detailed flow chart of a navigation/execution phase of the application of FIG. 4; and
FIGS. 8-27 are representative screens of graphical user interface pages displayed by the computer-assisted surgery system of FIG. 1 during use of the application of FIG. 4.
DETAILED DESCRIPTION OF THE DRAWINGS In the following description, like numbers refer to like elements. References to
"surgeon" include any user of a computer-assisted surgical system, a surgeon being typically a primary user.
FIG. 1 is a block diagram of an exemplary computer-assisted surgery (CAS) system
10. Computer-assisted surgery system (CAS) 10 comprises a display device 12, an input device 14, and a processor-based system 16, for example, a computer. Display device 12 may be any display device now known or later developed for displaying two-dimensional and/or three-dimensional diagnostic images, for example, a monitor, a touch screen, a wearable display, a projection display, a head-mounted display, stereoscopic views, a holographic display, a display device capable of displaying image(s) projected from an image projecting device, for example, a projector, and/or the like. Input device 14 may be any input device now known or later developed, for example, a keyboard, a mouse, a trackball, a trackable probe and/or the like. The processor-based system is preferably programmable and includes one or more processors 16a, working memory 16b for temporary program and data storage that will be used primarily by the processor, and storage for programs and data, preferably persistent, such as a disk drive. Removable media storage device 18 can also be used to store programs and/or transfer to or from the transfer programs.
Tracking system 22 continuously determines, or tracks, the position of one or more trackable markers disposed on, incorporated into, or inherently a part of surgical tools or instruments 20 with respect to a three-dimensional coordinate frame of reference. With information from the tracking system on the location of the trackable markers, CAS system 10 is programmed to be able to determine the three-dimensional coordinates of an endpoint or tip of a tool and, optionally, its primary axis using predefined or known (e.g. from calibration) geometrical relationships between trackable markers on the tool and the end point and/or axis of the tool. A patient, or portions of the patient's anatomy, can also be tracked by attachment of arrays of trackable markers.
The CAS system can be used for both planning surgical procedures (including planning during surgery) and for navigation. It is therefore preferably programmed with software for providing basic image-guided surgery functions, including those necessary determining the position of the tip and axis of instruments and for registering a patient and preoperative and/or intraoperative diagnostic image data sets to the coordinate system of the tracking system. The programmed instructions for these functions are indicated as core CAS utilities 24. These capabilities allow the relationship of a tracked instrument to a patient to be displayed and constantly updated in real time by the CAS system overlaying a representation of the tracked instrument on one or more graphical images of the patient's internal anatomy on display device 12. The graphical images are constructed from one or more stored image data sets 26 acquired from diagnostic imaging device 28. Imaging device may be a fluoroscope, such as a C-arm fluoroscope, capable of being positioned around a patient lying on an operating table. It may also be a MR, CT or other type of imaging device in the room or permanently located elsewhere. Where more than one image is shown, as when multiple fluoroscopic images are simultaneously displayed of display device 12, the representation of the tracked instrument or tool is coordinated between the different images. However, CAS system can be used in some procedures without the diagnostic image data sets, with only the patient being registered. Thus, the CAS system need not support the use of diagnostic images in some applications - i.e. an imageless application.
Furthermore, as disclosed herein, the CAS system may be used to run application- specific programs 30 that are directed to assisting a surgeon with planning and/or navigation during specific types of procedures. For example, the application programs may display predefined pages or images corresponding to specific steps or stages of a surgical procedure. At a particular stage or part of a program, a surgeon may be automatically prompted to perform certain tasks or to define or enter specific data that will permit, for example, the program to determine and display appropriate placement and alignment of instrumentation or implants or provide feedback to the surgeon. Other pages may be set up to display diagnostic images for navigation and to provide certain data that is calculated by the system for feedback to the surgeon. Instead of or in addition to using visual means, the CAS system could also communicate information in ways, including using audibly (e.g. using voice synthesis) and tactilely, such as by using a haptic interface of device. For example, in addition to indicating visually a trajectory for a drill or saw on the screen, a CAS system may feedback to a surgeon information whether he is nearing some object or is on course with a audible sound or by application of a force or other tactile sensation to the surgeon's hand.
To further reduce the burden on the surgeon, the program may automatically detect the stage of the procedure by recognizing/identifying the instrument picked up by a surgeon and move immediately to the part of the program in which that tool is used. Application data 32 - data generated or used by the application — may also be stored processor-based system.
Various types of user input methods can be used to improve ease of use of the CAS system during surgery. One example is the use of speech recognition to permit a doctor to speak a command. Another example is the use of a tracked object to sense a gesture by a surgeon, which is interpreted as an input to the CAS system. The meaning of the gesture could further depend on the state of the CAS system or the current step in an application process executing on the CAS system. Again, as an example, a gesture may instruct the CAS system to capture the current position of the object. One way of detecting a gesture is to occlude temporarily one or more of the trackable markers on the tracked object (e.g. a probe) for a period of time, causing loss of the CAS system's ability to track the object. A temporary visual occlusion of a certain length (or within a certain range of time), coupled with the tracked object being in the same position before the occlusion and after the occlusion, would be interpreted as an input gesture. A visual or audible indicator that a gesture has been recognized could be used to provide feedback to the surgeon.
Yet another example of such an input method is the use of tracking system 22 in combination with one or more trackable data input devices 34. Defined with respect to the trackable input device 34 are one or more defined input areas, which can be two-dimensional or three-dimensional. These defined input areas are visually indicated on the trackable input device so that a surgeon can see them. For example, the input areas may be visually defined on an object by representations of buttons, numbers, letters, words, slides and/or other conventional input devices. The geometric relationship between each defined input area and the trackable input device is known and stored in processor-based system 16. Thus, the processor can determine when another trackable object touches or is in close proximity a defined input area and recognize it as an indication of a user input to the processor-based systems. For example, when a tip of a tracked pointer is brought into close proximity to one of the defined input areas, the processor-based system will recognize the tool near the defined input area and treat it as a user input associated with that defined input area. Preferably, representations on the trackable user input correspond with user input selections (e.g. buttons) on a graphical user interface on display device 12. The trackable input device may be formed on the surface of any type of trackable device, including devices used for other purposes. In a preferred embodiment, representations of user input functions for graphical user interface are visually defined on a rear, flat surface of a base of a tool calibrator.
Processor-based system 16 is, in one example, a programmable computer that is programmed to execute only when single-use or multiple-use software is loaded from, for example, removable media. The software would include, for example the application program 30 for use with a specific type of procedure. Media storing the application program can be sold bundled with disposable instruments specifically intended for the procedure. The application program would be loaded into the processor-based system and stored there for use during one (or a defined number) of procedures before being disabled. Thus, the application program need not be distributed with the CAS system. Furthermore, application programs can be designed to work with specific tools and implants and distributed with those tools and implants. Preferably, also, the most current core CAS utilities may also be stored with the application program. If the core CAS utilities on the processor-based system are outdated, they can be replaced with the most current utilities. FIG. 2 is intended to be a representative patient with a representative fractured femur. The representative patient 200, represented by a head 202, torso 204, arm 206, leg 208 and knee 210. Indicated by dashed lines in the upper leg, above the knee, is a femur 212 that is fractured and separated into two pieces, which will be referred to as the proximal fragment 214 and distal fragment 216 to correspond with the proximal end 218 and distal end 220 of the femur. A trackable marker array 212, which can be tracked by the CAS system 10 (FIG. 1), is attached to, respectively, the proximal piece 214 and distal piece 216 of the femur so that the relative position of the two pieces can be tracked during implantation of an TM nail into the femur. Referring now to FIG. 3, the CAS system assists a surgeon in performing an BVI nail implantation by executing a process 300 that has three basic phases: set-up phase 302, planning phase 304 and navigation phase 306. The set-up phase involves the surgeon specifying to the process which type of BVI nail to be used which leg is to be operated on, type of fracture, instruments and/or tools to be tracked during the procedure, and model fluoroscope to be used, which leg is to be operated on, type of fracture, instruments, and/or tools during the process and certain other options such as determining to image and plan using the uninjured leg. The set-up phase allows for skipping certain steps during the navigation or execution stage so that it flows more efficiently to the surgeon's preferences or needs. The planning phase involves using fluoroscopic images to gather reference information on leg version (rotation angle) and length from the surgeon and to select nail dimensions, and placement and length of screws used to secure the nail. The navigation or execution stage tracks the surgeon's instruments and trackable markers implanted in or attached to the patient's femur and provides alignment information and feedback on version and length. Process 300, or parts thereof, preferably display a series of pages corresponding to stages or sub-procedures, each page being set up to display directions and information (including images) relevant to the stage of the procedure. However, as previously mentioned, the CAS system may in addition to the pages or in place of the pages, communicate some or all of this information by other means, including audible and haptic means. Although the process may constrain what a surgeon does in terms of the ordering of certain steps, the process preferably follows the surgeon, rather than requiring the surgeon to follow the process. This is particularly useful during the planning and navigation or execution phases of the process, where the surgeon may need to go back and change a plan or repeat steps. Thus, in the following explanation of process 300, some steps maybe performed out of sequence or repeated. The surgeon may indicate to the process the stage he or she is in or wants to go to. This may be done through user input or by the process automatically recognizing when the surgeon has either finished a stage or is preparing to go to another stage (not necessarily the next stage) by, for example, the surgeon picking up an instrument used in a particular stage and showing it to the cameras of the tracking system. Details of the process 300 will be described with reference to representative examples of screens from such pages, shown in FIGS. 8-27. These screens contemplate use of IM nails for a specific vendor. However, the process and concepts embodied or represented by the pages are not limited to any specific vendor, and aspects thereof may be employed in connection with surgical planning and guidance systems for similar types of implants.
Referring to FIG. 4 and FIGS 8-13, step 402 asks the surgeon to identify or select which of a plurality of IM nail types or families will be used. This information is used for representing the LM nail on images taken of the patient's leg and providing feedback to the surgeon on the position of the nail during the nail insertion. If the process is set up for only one type of nail, this step may be skipped. FIG. 8 is a screen of a representative example of such a page, in this case showing four families of IM nails from a particular vendor. At steps 404, 406, and 408, the process requests the surgeon to specify which leg is injured, what type of fluoroscope will be used, and whether the uninjured leg of the patient will be used in planning. FIG. 9 is an example of a graphical interface displaying the options for selection by the surgeon. Although the use a fluoroscopic images has certain advantages, other types of images can be used in place of, or in addition to, the fluoroscopic images, including without limitation preoperative three-dimensional data sets such as CT and MRI scans.
At step 410 the surgeon is asked to specify application-specific tools that he will use during the procedure that can be or will be tracked. Surgeons may prefer to use different tools for a given step, and this step permits the surgeon to select the tool of choice so that the CAS system can properly track it. The application may display a different page at a given step, or display pages in a different order, based on selection of the tool. Furthermore, a surgeon may, for example, elect not to use a tool during a given step, or not have it tracked. The process will adjust as necessary to accommodate the preferences to avoid forcing a surgeon to find ways to bypass steps or alter presentation of the pages. The CAS system is typically programmed or set up to operate with a probe and other basic tools that a surgeon may use. Preferably, the surgeon is given a list of the tool or tools that the application can track, from which he may select. FIG. 10 shows an example of a page that displays the tools that the application is capable of or set up to track for the basic steps of the surgical procedure. The display permits the surgeon to visually select the tool (or not to have a tool) and verify the selection. The tool listing in the illustrated example is also grouped by basic stages of the process. In the example, options are given for the tool that will be used for defining an entry point in the femur prior to insertion of the nail, the tool used for nail insertion, the instrument used for drilling holes to insert screws for locking the distal end of the nail, and other tools that the surgeon may want to use. Thus, in the example, if no tool is selected for specifying the entry point, the program will not expect to receive an indication for the entry point and will not attempt to display the selected point on a diagnostic image. If, for example, a surgeon selects a power drill instead of a hand drill for distal locking, the CAS system will automatically assume that it is tracking a power drill during the distal locking step.
At step 412, the CAS system calibrates the selected fluoroscope using known methods. The interface for this step is illustrated in FIG. 11.
Steps 414, 416, 418 and 420 direct the acquisition of certain fluoroscopic images during the procedure, followed by registration of those images using known methods. If the surgeon specified that the well leg would be used for reference, images of the well leg are acquired in addition to images of the injured leg. Exemplary screen shots of the pages corresponding to the acquisition and registration of the well leg and injured leg are shown in FIGS. 12 and 13, respectively. What images are needed or desirable are listed and identified with respect to the list when they are acquired. In the illustrated examples, the required or desirable images are listed with reference to target areas 1202 defined on diagrams 1204 of a femur. The diagrams show a femur from an anterior/posterior and from a medial/lateral view. It is preferred to acquire an anterior/posterior (A/P) and a medial/lateral (M/L) image of each of the proximal and distal ends of the femur, fluoroscopic image pair is around the midshaft of the femur. Area 1202 may consist of two fracture sites, depending upon the setup phase. This allows for handling a compound fracture and allows the surgeon to image the proximal and distal fractures in separate shots. Each A/P and M/L fluoroscopic image pair is preferably shown in two, side-by-side windows on the display. During image acquisition, window 1206 displays a current image from the fluoroscope. Once a surgeon is satisfied with an image, it is saved or stored by the CAS system upon appropriate input from the surgeon and is moved to adjacent window 1208 for registration. Referring now to FIG. 5, if a well leg is imaged, the planning stage starts with a process 500. With the images of the well leg, certain reference information, namely a reference length and version, are determined at step 502, based on information indicated on the images by the surgeon. Assuming that the injured and well legs are anatomically similar, the well leg may also be used to at least initially determine appropriate nail length and diameter at step 504. Using the well leg to determine this information may be desirable in the event it is difficult to determine this information from the injured leg. As indicated by step 506, screw length placement and length may also be determined.
FIGS. 6A and 6B illustrate in the planning stage in greater detail. Steps 602 to 608 involve determination of a reference length and version of the leg or femur. During these steps the surgeon is prompted to indicate in the acquired images of the well leg certain anatomical landmarks, preferably the center of the femoral head, the axis of the femoral neck and shaft, an axis that extends transverse to the condyles at the posterior-most points of the condyles (the trans-epicondylar axis). However, other recognizable landmarks could be used for calculating a reference length and/or version.
At steps 601 and 602 the surgeon is prompted to indicate, and the process receives, an estimated nail diameter on isthmus of uninjured leg and the center of the femoral head and the axis of its neck with reference to displayed A/P and M/L images of the proximal end of the femur. As illustrated in representative page or interface of FIG. 14, a "bull's-eye" marker 1402 is superimposed on A/P image 1404 and M/L image 1406 for assisting the surgeon in identifying the center of the generally spherical femoral head in both images. This bull's-eye marker is a two-dimensional projection of a series of nested, virtual spheres in the three dimensional space of the patient. As the surgeon moves the marker with respect to one image, its position is automatically updated with respect to the second image. The surgeon is, in effect, moving the virtual spheres. From the center point of the bull's-eye marker to a second end point extends another marker in the form of line 1408. It represents a virtual guide wire in the three-dimensional space of the patient. The surgeon moves this virtual guide wire so that it extends along the axis of the femoral neck. Once the surgeon indicates that markers are in the correct position, the process moves automatically to step 604. At step 604, the process displays A/P and lateral images of the distal end of the femur.
The surgeon indicates on the images a marker for services as a reference point for determining a reference length for the femur. The program stores this information. At step 606, the reference length is calculated using the references marked on the proximal end of the femur and the reference marked on the distal end of the femur. The program also prompts, using, for example, directions displayed on the displayed page, and receives from the surgeon at step 406 the position and orientation for the trans-epicondylar axis of the femur. FIG. 15 is an example of a screen used in these steps. The acquired A P and lateral images, 1502 and 1504, respectively, of the distal end of the well femur are displayed, along with a reference line 1506. The surgeon manipulates the position and orientation of the reference line in A/P image 1502 so that it is aligned with the trans-epicondylar axis. In the lateral view the surgeon manipulates the reference line 1506 so that it is positioned on the posterior most points of the condyles. Using the definition of the femoral neck axis received at step 604 and the definition of the trans-epicondylar axis received at step 406, a reference version is calculated at step 608. The trans-epicondylar axis also serves as a reference point for calculating a reference length. In order to have meaningful version and reference information, true A/P and lateral views of the distal end of the femur should be acquired, or they should be least taken at the substantially same angles as the A P and lateral images of the distal end of the injured femur. The reference length and version are stored and displayed in area 1508 of the screen. As an alternative to using a reference line, the version may be calculated using a true lateral image of the distal end and placing a reference point on the knee center in both the A/P and lateral images.
Steps 608, 610, 611, 612, 614 and 616 assist the surgeon with selecting a nail of appropriate length and screw dimensions using the well leg. At step 608, the surgeon indicates, with respect to A/P and M/L images of the distal and proximal ends of the femur, end points for the nail. The process automatically determines the distance between the end points and then it selects and displays on the images a representation of the closest standard length nail. As indicated by steps 610 and 611, screw placement and dimensions for the proximal end of the nail and the placement of the nail end are indicated with respect to the uninjured leg. A representation of the closest standard nail to the indications is then displayed at step 612. The surgeon is then permitted to change, shift, rotate and move the representation in order to check its fit. If the fit is not correct, the surgeon can change the end points and/or select a different nail, as indicated by steps 614 and 616. FIG. 16 is a representative screen from an example of a user interface page displayed on the CAS system implementing these steps. The page includes the stored A/P image 1602 and M/L image 1604 of the proximal end of the well-leg femur. Superimposed on this image is a marker, in the form of a crosshair graphic 1606, for marking the estimated proximal end of the nail that will be implanted in the other (injured) leg. A representation 1608 of the proximal end of the nail is also preferably superimposed on the two images, along with representations 1610 of screws that will be inserted through the proximal end of the femur and nail once the nail is fully inserted. The surgeon is permitted to change, shift, rotate and move the representation of the screws in order to check its fit. The page includes inputs for changing the position of the cross-hairs and representations. FIG. 17 is a representative screen of a page for the surgeon to mark an estimated location for the tip of the nail. The page includes the stored A/P image 1602 and M/L image 1606 of the distal end of the well- leg femur. It prompts the surgeon to move a marker, namely, the cross-hairs graphic 1706, to the estimated tip of the nail. The program then provides an estimated nail length and displays the two closest standard lengths for the type of nail being used on line 1710. The surgeon selects the desired length and the program moves the nail representation to the correct length for the surgeon to confirm that selected length.
Referring now to FIG. 6B, the diameter of the nail is estimated using at step 618, the midshaft (isthmus) of the well or injured leg. FIG. 18 assumes that well-leg images were acquired at the midshaft of that leg's femur. The diameter of the canal of the femur, through which the nail will be inserted, is its narrowest at the isthmus. The page instructs the surgeon to place a reference marker 1806 along the canal of the femur and then select the best matching diameter from a list. As different nail diameters are selected, the width of the reference marker, which is a projection of a virtual, cylindrical object (corresponding generally to a diameter of a nail) in the three-dimensional patient space into the two- dimensional fluoroscopic images, changes. Once the surgeon decides on a diameter, it is stored and the process moves to injured leg planning.
At step 620, the surgeon is prompted to mark in the images showing the edge of the fracture at the canal of the femur. A representative screen of the page displayed for this step is shown in FIG. 19. The stored A/P image 1902 and the stored lateral M/L image 1904 of the midshaft of the injured femur is displayed, and cross-hairs marker 1906 is also displayed and can be moved by the surgeon to mark the edge of the fracture. A representation 1908 of the nail is also superimposed for the surgeon to check and, if necessary, change the estimated nail diameter that will fit through the canal at the point of fracture.
Injured-leg planning continues at steps 622 at the proximal end of the injured femur by the surgeon marking in the images the center of the femoral head and the axis of the femoral neck substantially in the same manner as discussed in connection with step 602. This information will be used to calculate reference length and version for the injured leg. FIG. 20 is a representative screen of the page displayed for this step. Its display includes the stored A/P and M/L images of the proximal end of the femur in windows 2002 and 2004. Like other pages, it includes written instructions prompting the surgeon to mark certain landmarks, namely, the femoral head and neck using a bull's eye marker 2006 and 2008, just as in FIG. 14.
In a manner similar to step 606, the same landmarks used in marking the distal end of the well femur in step 606 are marked at step 624 by the surgeon and stored for use in calculating reference length and version for comparison to the well leg. A representative screen of a display page for this step is shown in FIG. 21. It includes the stored A/P and M/L images 2102 and 2104 for the distal end of the injured leg. It also includes an A/P shot of the distal end of the well femur 2106 for reference to ensure proper marking of the landmarks on the images 2102 and 2104. The reference line is shown on image 2106 in the position marked by the surgeon at step 606. As with step 606, trans-epicondylar axis is marked on the images with reference line 2108 and stored.
Once the reference points are marked, the process proceeds to steps 628 and 630, where the surgeon indicates to the process the entry point for the nail and desired position of the nail head and the screws that lock the nail head. As show in FIG. 22, a representative screen of an example of a page for receiving this information from the surgeon, the stored A/P and M/L images 2202 and 2204 of the distal end of the injured femur are displayed and overlaid with a representation 2206 of the previously selected nail and the locking screws 2208. The nail head 2210, which defines the entry point for the nail into the femur, is also indicated. The surgeon shifts and rotates the representation of the nail so that it fits properly in the canal and the locking screws extend up the neck of the femur shown in the images. The representations of the screws are fixed to the representation of the nail, and rotate and shift with it. When the surgeon is satisfied with the placement of the nail and locking screws, this information is stored.
As a final step before execution, tools previously selected for use in the procedure are calibrated if they are not already calibrated at step 632. A representative screen of an exemplary page that may be displayed at this step is shown in FIG. 23. A list 2302 of selected tools is displayed. A surgeon selects each tool on the list for calibration. When the tool comes into the field view of the tracking system of the CAS system, the tool is recognized and instructions for calibration are displayed. During this step, the tip and, optionally, axis of each tool is calculated with respect to a known point on a calibration fixture according to known methods. The calibration information is stored by the CAS system so that the relationship between the displayed representation of the tool and the diagnostic images is the same as the relationship between the actual tool and the patient. Referring now to FIG. 7, steps 702, 704, 706 and 708 involve guiding the surgeon to the correct entry point for inserting the nail. Referring now also to FIG. 24, the previously acquired and stored A P and M/L images 2402 and 2404 are displayed. The entry point is also marked with markers 2406. Although not explicitly shown in the figures, the point of the tool selected/or use in forming the entry, in this case an awl, is continuously tracked by the CAS system and a representation of the position of the tip of the tool displayed on the images. The CAS system is continuous tracking relative changes in positions of the two fragments using trackable marker array 224 (see FIG. 2) attached to each fragment. The arrays are attached prior to registration of the images with the patient so that registration is not lost due to movement of either femoral fragment. With each leg fragment being tracked to maintain registration, the CAS system will compensate for movement of the fragments when displaying the position of tracked tools on the diagnostic images on the CAS system's display. In area or window 2414 of the display, illustrations of the objects being tracked are displayed, in this example, and awl and a trackable marker array. These illustrations also provide an indication to the surgeon if the tool or the marker array are out of the field of view of the camera by displaying, in the illustrated embodiment, a red outline on respective images in area 2414. This function is also present in subsequent tracking steps.
In window or area 2407 the relative positions and orientations of the proximal and femoral fragments of the fractured femur are indicated by representations 2410 and 2408. This window is preferably displayed during steps 708 and 714. Displayed in area 2412 is reference length and version infonnation that is continuously calculated based on the relative positions of the fragments. This tracking is possible due to the known relationship between trackable marker array and the reference landmarks specified on the fragment. At the time when the landmarks on each fragment were specified, the positions of the trackable markers were also stored, thereby permitting the relative relationship to be determined. Using the relative relationship between each trackable marker 224 and the landmarks on the fragment to which it is attached, the referenced lengths and version are calculated based on the relative positions of the two trackable markers. Referring now to FIGS 7 and FIG. 25, after the surgeon forms the entry for the nail, he will "ream" the canal of the femur to prepare it for introduction of the nail. Although not tracked in the illustrated embodiment, the instrument used for reaming could be tracked to and its position could also be displayed on the images to ensure that it successfully bridges the fracture and enters the canal of the other fragment. Since the reaming device must be flexible and is located inside the femur, optical tracking cannot be used. Magnetic tracking, though less precise, could be employed. Once the canal is prepared, the surgeon will employ a tool for inserting the nail, referred to as a nail inserter. The inserter is tracked by the CAS system. Bringing the nail inserter into view of the tracking system signals the application process to move to the next step, namely, to step 710. The geometric relationship between the tool and the nail is known from the calibration step performed earlier. Therefore, by tracking the tool inserter, which remains outside the patient, the position of the nail is known.
In FIG. 25, the stored A/P and M/L images of the proximal end of the femur are displayed. Also displayed on images using representation 2506 is the current position of the nail and the screws as the nail is being inserted and rotated. The nail insertion tool is tracked. The position of the nail and screws is determined from the position of the nail insertion tool and the geometric relationship between the nail insertion tool and nail. As in FIG. 24, window 2407 displays the representations of the two fragments 2408 and 2410, of the femur in the relative positions and calculated reference lengths and versions 2412. The surgeon will use the nail and screw representations to ensure that the screws are correctly aligned with the femoral neck. The representations of the locking screws can be used as guides for drilling and inserting the screws.
Once the surgeon inserts the nail and the proximal locking screws, the distal end locking screws must be inserted. The nail guide does not typically incorporate an external guide due at least in part to a possibility of the nail bending during insertion. In order to locate screw openings in the nail and determine trajectory of the screws, another set of lateral and A/P and M/L images of the distal end of the femur is required. Therefore, at step 712, the surgeon is prompted to acquire the additional images. FIG. 26 is an example of a page for guiding the surgeon in capturing the images. The current image for the fluoroscope is shown in window 2602. If the image is acceptable, it is stored and shown in window 2604. The shots or images to be acquired are, in this example, graphically illustrated in area 2606.
The second set of stored A/P and M/L images of the distal end of the femur should clearly show the screw holes in the distal end of the nail. In order to clearly see the holes, the lateral image needs to be a true lateral image relative to the nail. When a surgeon brings the instrument previously specified as being used for distal screw insertion into the area of focus of the tracking system, the CAS system preferably automatically displays a screen or page similar to the one of FIG. 27 and performs steps 714 and 716. The page of FIG. 27 includes the stored A/P image 2702 and lateral image 2704 of the distal end of the nail. To guide a surgeon in inserting the locking screw, a representation 2706 of the instrument being used for the insertion is superimposed on the images. A representation 2708 of the locking screw on the end of the instrument is also superimposed.
At the conclusion of the procedure, the surgeon is prompted to specify whether to archive data generated by the procedure for later reference. The CAS system archives the data as directed, such as to a disk drive or removable media. This step is not illustrated.
If desired, the different steps discussed herein may be performed in any order and/or concurrently with each other. Furthermore, if desired, one or more of the above described steps may be optional or may be combined without departing from the scope of the present invention.
Embodiments of the present invention may be implemented in software, hardware, application logic or a combination of software, hardware and application logic. The software, application logic and/or hardware may reside on processor-based system 16 or on a removable storage medium. If desired, part of the software, application logic and/or hardware may reside on processor-based system 16 and part of the software, application logic and/or hardware may reside on the removable storage medium.

Claims

ClaimsWhat is claimed is:
1. Apparatus for assisting with surgical procedure, comprising: a localizer; a computer in communication with the localizer, the computer storing and executing instructions for displaying a plurality of screens, a first one of the plurality of screens corresponding to a planning step for a procedure for inserting an intramedullary nail and a second one of the plurality of screens corresponding to a navigation step of the procedure, the first one of the plurality of screens assisting with selection of the nail based on a patients anatomy and the second one of the plurality of screens indicating the position of the nail as it is being inserted into the patient's femur.
PCT/US2004/003068 2003-02-04 2004-02-04 Method and apparatus for computer assistance with intramedullary nail procedure WO2004069040A2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US44500103P 2003-02-04 2003-02-04
US60/445,001 2003-02-04

Publications (2)

Publication Number Publication Date
WO2004069040A2 true WO2004069040A2 (en) 2004-08-19
WO2004069040A3 WO2004069040A3 (en) 2005-03-24

Family

ID=32850960

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2004/003068 WO2004069040A2 (en) 2003-02-04 2004-02-04 Method and apparatus for computer assistance with intramedullary nail procedure

Country Status (2)

Country Link
US (2) US20060173293A1 (en)
WO (1) WO2004069040A2 (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2191783A1 (en) * 2008-01-09 2010-06-02 Stryker Leibinger GmbH & Co. KG Stereotactic computer assisted surgery based on three-dimensional visualization
US7840256B2 (en) 2005-06-27 2010-11-23 Biomet Manufacturing Corporation Image guided tracking array and method
US8010180B2 (en) 2002-03-06 2011-08-30 Mako Surgical Corp. Haptic guidance system and method
US8287522B2 (en) 2006-05-19 2012-10-16 Mako Surgical Corp. Method and apparatus for controlling a haptic device
WO2013174401A1 (en) * 2012-05-23 2013-11-28 Stryker Trauma Gmbh Entry portal navigation
US9508149B2 (en) 2012-05-23 2016-11-29 Stryker European Holdings I, Llc Virtual 3D overlay as reduction aid for complex fractures
US9517107B2 (en) 2010-07-16 2016-12-13 Stryker European Holdings I, Llc Surgical targeting system and method
US9801686B2 (en) 2003-03-06 2017-10-31 Mako Surgical Corp. Neural monitor-based dynamic haptics
US9855104B2 (en) 2012-05-23 2018-01-02 Stryker European Holdings I, Llc Locking screw length measurement
US10039606B2 (en) 2012-09-27 2018-08-07 Stryker European Holdings I, Llc Rotational position determination
EP3241166A4 (en) * 2014-12-31 2018-10-03 Vector Medical, LLC Process and apparatus for managing medical device selection and implantation
US10588647B2 (en) 2010-03-01 2020-03-17 Stryker European Holdings I, Llc Computer assisted surgery system
US10610301B2 (en) 2002-03-06 2020-04-07 Mako Surgical Corp. System and method for using a haptic device as an input device
US11202676B2 (en) 2002-03-06 2021-12-21 Mako Surgical Corp. Neural monitor-based dynamic haptics

Families Citing this family (105)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050277823A1 (en) * 2002-06-10 2005-12-15 Robert Sutherland Angiogram display overlay technique for tracking vascular intervention sites
WO2004014219A2 (en) * 2002-08-09 2004-02-19 Kinamed, Inc. Non-imaging tracking tools and method for hip replacement surgery
US8583220B2 (en) * 2005-08-02 2013-11-12 Biosense Webster, Inc. Standardization of catheter-based treatment for atrial fibrillation
US7877128B2 (en) * 2005-08-02 2011-01-25 Biosense Webster, Inc. Simulation of invasive procedures
US8219178B2 (en) 2007-02-16 2012-07-10 Catholic Healthcare West Method and system for performing invasive medical procedures using a surgical robot
US10653497B2 (en) 2006-02-16 2020-05-19 Globus Medical, Inc. Surgical tool systems and methods
US10357184B2 (en) 2012-06-21 2019-07-23 Globus Medical, Inc. Surgical tool systems and method
US10893912B2 (en) 2006-02-16 2021-01-19 Globus Medical Inc. Surgical tool systems and methods
EP2051647B1 (en) * 2006-08-15 2010-12-01 AO Technology AG Method and device for computer assisted distal locking of intramedullary nails
US8934961B2 (en) 2007-05-18 2015-01-13 Biomet Manufacturing, Llc Trackable diagnostic scope apparatus and methods of use
JP5250251B2 (en) * 2007-12-17 2013-07-31 イマグノーシス株式会社 Medical imaging marker and its utilization program
EP2286370A4 (en) * 2008-05-02 2014-12-10 Eyeic Inc System for using image alignment to map objects across disparate images
FR2937530B1 (en) * 2008-10-24 2012-02-24 Biospace Med MEASURING INTRINSIC GEOMETRIC SIZES WITH AN ANATOMIC SYSTEM
CN102316799A (en) * 2009-02-11 2012-01-11 皇家飞利浦电子股份有限公司 Method and system of tracking and mapping in a medical procedure
US8786873B2 (en) * 2009-07-20 2014-07-22 General Electric Company Application server for use with a modular imaging system
WO2012131660A1 (en) 2011-04-01 2012-10-04 Ecole Polytechnique Federale De Lausanne (Epfl) Robotic system for spinal and other surgeries
CN103733200B (en) * 2011-06-27 2017-12-26 皇家飞利浦有限公司 Checked by the inspection promoted with anatomic landmarks clinical management
US11045267B2 (en) 2012-06-21 2021-06-29 Globus Medical, Inc. Surgical robotic automation with tracking markers
US11793570B2 (en) 2012-06-21 2023-10-24 Globus Medical Inc. Surgical robotic automation with tracking markers
US11395706B2 (en) 2012-06-21 2022-07-26 Globus Medical Inc. Surgical robot platform
US11253327B2 (en) 2012-06-21 2022-02-22 Globus Medical, Inc. Systems and methods for automatically changing an end-effector on a surgical robot
US11298196B2 (en) 2012-06-21 2022-04-12 Globus Medical Inc. Surgical robotic automation with tracking markers and controlled tool advancement
US10136954B2 (en) 2012-06-21 2018-11-27 Globus Medical, Inc. Surgical tool systems and method
US10231791B2 (en) 2012-06-21 2019-03-19 Globus Medical, Inc. Infrared signal based position recognition system for use with a robot-assisted surgery
US11399900B2 (en) 2012-06-21 2022-08-02 Globus Medical, Inc. Robotic systems providing co-registration using natural fiducials and related methods
US10350013B2 (en) 2012-06-21 2019-07-16 Globus Medical, Inc. Surgical tool systems and methods
US11864839B2 (en) 2012-06-21 2024-01-09 Globus Medical Inc. Methods of adjusting a virtual implant and related surgical navigation systems
US10758315B2 (en) 2012-06-21 2020-09-01 Globus Medical Inc. Method and system for improving 2D-3D registration convergence
US11864745B2 (en) 2012-06-21 2024-01-09 Globus Medical, Inc. Surgical robotic system with retractor
US11607149B2 (en) 2012-06-21 2023-03-21 Globus Medical Inc. Surgical tool systems and method
US11116576B2 (en) 2012-06-21 2021-09-14 Globus Medical Inc. Dynamic reference arrays and methods of use
US11317971B2 (en) 2012-06-21 2022-05-03 Globus Medical, Inc. Systems and methods related to robotic guidance in surgery
JP2015528713A (en) 2012-06-21 2015-10-01 グローバス メディカル インコーポレイティッド Surgical robot platform
US10624710B2 (en) 2012-06-21 2020-04-21 Globus Medical, Inc. System and method for measuring depth of instrumentation
US11857266B2 (en) 2012-06-21 2024-01-02 Globus Medical, Inc. System for a surveillance marker in robotic-assisted surgery
US11857149B2 (en) 2012-06-21 2024-01-02 Globus Medical, Inc. Surgical robotic systems with target trajectory deviation monitoring and related methods
CA2882388C (en) 2012-08-31 2020-10-20 Sloan-Kettering Institute For Cancer Research Particles, methods and uses thereof
EP2958481A4 (en) * 2013-02-20 2017-03-08 Sloan-Kettering Institute for Cancer Research Wide field raman imaging apparatus and associated methods
DE102013210185A1 (en) * 2013-05-31 2014-12-04 Siemens Aktiengesellschaft Method of visual assistance in fixing an implant
US9283048B2 (en) 2013-10-04 2016-03-15 KB Medical SA Apparatus and systems for precise guidance of surgical tools
WO2015107099A1 (en) 2014-01-15 2015-07-23 KB Medical SA Notched apparatus for guidance of an insertable instrument along an axis during spinal surgery
US10039605B2 (en) 2014-02-11 2018-08-07 Globus Medical, Inc. Sterile handle for controlling a robotic surgical system from a sterile field
US10912947B2 (en) 2014-03-04 2021-02-09 Memorial Sloan Kettering Cancer Center Systems and methods for treatment of disease via application of mechanical force by controlled rotation of nanoparticles inside cells
WO2015162256A1 (en) 2014-04-24 2015-10-29 KB Medical SA Surgical instrument holder for use with a robotic surgical system
US10357257B2 (en) 2014-07-14 2019-07-23 KB Medical SA Anti-skid surgical instrument for use in preparing holes in bone tissue
CA2956210A1 (en) 2014-07-28 2016-02-04 Memorial Sloan Kettering Cancer Center Metal(loid) chalcogen nanoparticles as universal binders for medical isotopes
US10013808B2 (en) 2015-02-03 2018-07-03 Globus Medical, Inc. Surgeon head-mounted display apparatuses
WO2016131903A1 (en) 2015-02-18 2016-08-25 KB Medical SA Systems and methods for performing minimally invasive spinal surgery with a robotic surgical system using a percutaneous technique
EP3317035A1 (en) 2015-07-01 2018-05-09 Memorial Sloan Kettering Cancer Center Anisotropic particles, methods and uses thereof
US10646298B2 (en) 2015-07-31 2020-05-12 Globus Medical, Inc. Robot arm and methods of use
US10058394B2 (en) 2015-07-31 2018-08-28 Globus Medical, Inc. Robot arm and methods of use
US10080615B2 (en) 2015-08-12 2018-09-25 Globus Medical, Inc. Devices and methods for temporary mounting of parts to bone
JP6894431B2 (en) 2015-08-31 2021-06-30 ケービー メディカル エスアー Robotic surgical system and method
US10034716B2 (en) 2015-09-14 2018-07-31 Globus Medical, Inc. Surgical robotic systems and methods thereof
US9771092B2 (en) 2015-10-13 2017-09-26 Globus Medical, Inc. Stabilizer wheel assembly and methods of use
US11058378B2 (en) 2016-02-03 2021-07-13 Globus Medical, Inc. Portable medical imaging system
US10842453B2 (en) 2016-02-03 2020-11-24 Globus Medical, Inc. Portable medical imaging system
US10448910B2 (en) 2016-02-03 2019-10-22 Globus Medical, Inc. Portable medical imaging system
US11883217B2 (en) 2016-02-03 2024-01-30 Globus Medical, Inc. Portable medical imaging system and method
US10117632B2 (en) 2016-02-03 2018-11-06 Globus Medical, Inc. Portable medical imaging system with beam scanning collimator
US10866119B2 (en) 2016-03-14 2020-12-15 Globus Medical, Inc. Metal detector for detecting insertion of a surgical device into a hollow tube
SE542045C2 (en) * 2016-05-15 2020-02-18 Ortoma Ab Calibration object, system, and method calibrating location of instrument in a navigation system
JP7233841B2 (en) 2017-01-18 2023-03-07 ケービー メディカル エスアー Robotic Navigation for Robotic Surgical Systems
US11071594B2 (en) 2017-03-16 2021-07-27 KB Medical SA Robotic navigation of robotic surgical systems
US10390891B2 (en) 2017-06-13 2019-08-27 Biosense Webster (Israel) Ltd. Hologram lens for positioning an orthopedic implant
US10675094B2 (en) 2017-07-21 2020-06-09 Globus Medical Inc. Robot surgical platform
EP3492032B1 (en) 2017-11-09 2023-01-04 Globus Medical, Inc. Surgical robotic systems for bending surgical rods
US11357548B2 (en) 2017-11-09 2022-06-14 Globus Medical, Inc. Robotic rod benders and related mechanical and motor housings
US11794338B2 (en) 2017-11-09 2023-10-24 Globus Medical Inc. Robotic rod benders and related mechanical and motor housings
US11134862B2 (en) 2017-11-10 2021-10-05 Globus Medical, Inc. Methods of selecting surgical implants and related devices
US20190254753A1 (en) 2018-02-19 2019-08-22 Globus Medical, Inc. Augmented reality navigation systems for use with robotic surgical systems and methods of their use
US10573023B2 (en) 2018-04-09 2020-02-25 Globus Medical, Inc. Predictive visualization of medical imaging scanner component movement
US11337742B2 (en) 2018-11-05 2022-05-24 Globus Medical Inc Compliant orthopedic driver
US11278360B2 (en) 2018-11-16 2022-03-22 Globus Medical, Inc. End-effectors for surgical robotic systems having sealed optical components
US11744655B2 (en) 2018-12-04 2023-09-05 Globus Medical, Inc. Drill guide fixtures, cranial insertion fixtures, and related methods and robotic systems
US11602402B2 (en) 2018-12-04 2023-03-14 Globus Medical, Inc. Drill guide fixtures, cranial insertion fixtures, and related methods and robotic systems
US11382549B2 (en) 2019-03-22 2022-07-12 Globus Medical, Inc. System for neuronavigation registration and robotic trajectory guidance, and related methods and devices
US11806084B2 (en) 2019-03-22 2023-11-07 Globus Medical, Inc. System for neuronavigation registration and robotic trajectory guidance, and related methods and devices
US11419616B2 (en) 2019-03-22 2022-08-23 Globus Medical, Inc. System for neuronavigation registration and robotic trajectory guidance, robotic surgery, and related methods and devices
US11317978B2 (en) 2019-03-22 2022-05-03 Globus Medical, Inc. System for neuronavigation registration and robotic trajectory guidance, robotic surgery, and related methods and devices
US11571265B2 (en) 2019-03-22 2023-02-07 Globus Medical Inc. System for neuronavigation registration and robotic trajectory guidance, robotic surgery, and related methods and devices
US11045179B2 (en) 2019-05-20 2021-06-29 Global Medical Inc Robot-mounted retractor system
US11628023B2 (en) 2019-07-10 2023-04-18 Globus Medical, Inc. Robotic navigational system for interbody implants
US11571171B2 (en) 2019-09-24 2023-02-07 Globus Medical, Inc. Compound curve cable chain
US11426178B2 (en) 2019-09-27 2022-08-30 Globus Medical Inc. Systems and methods for navigating a pin guide driver
US11890066B2 (en) 2019-09-30 2024-02-06 Globus Medical, Inc Surgical robot with passive end effector
US11864857B2 (en) 2019-09-27 2024-01-09 Globus Medical, Inc. Surgical robot with passive end effector
US11510684B2 (en) 2019-10-14 2022-11-29 Globus Medical, Inc. Rotary motion passive end effector for surgical robots in orthopedic surgeries
US11382699B2 (en) 2020-02-10 2022-07-12 Globus Medical Inc. Extended reality visualization of optical tool tracking volume for computer assisted navigation in surgery
US11207150B2 (en) 2020-02-19 2021-12-28 Globus Medical, Inc. Displaying a virtual model of a planned instrument attachment to ensure correct selection of physical instrument attachment
US11253216B2 (en) 2020-04-28 2022-02-22 Globus Medical Inc. Fixtures for fluoroscopic imaging systems and related navigation systems and methods
US11382700B2 (en) 2020-05-08 2022-07-12 Globus Medical Inc. Extended reality headset tool tracking and control
US11510750B2 (en) 2020-05-08 2022-11-29 Globus Medical, Inc. Leveraging two-dimensional digital imaging and communication in medicine imagery in three-dimensional extended reality applications
US11153555B1 (en) 2020-05-08 2021-10-19 Globus Medical Inc. Extended reality headset camera system for computer assisted navigation in surgery
US11317973B2 (en) 2020-06-09 2022-05-03 Globus Medical, Inc. Camera tracking bar for computer assisted navigation during surgery
US11382713B2 (en) 2020-06-16 2022-07-12 Globus Medical, Inc. Navigated surgical system with eye to XR headset display calibration
US11877807B2 (en) 2020-07-10 2024-01-23 Globus Medical, Inc Instruments for navigated orthopedic surgeries
US11793588B2 (en) 2020-07-23 2023-10-24 Globus Medical, Inc. Sterile draping of robotic arms
US11737831B2 (en) 2020-09-02 2023-08-29 Globus Medical Inc. Surgical object tracking template generation for computer assisted navigation during surgical procedure
US11523785B2 (en) 2020-09-24 2022-12-13 Globus Medical, Inc. Increased cone beam computed tomography volume length without requiring stitching or longitudinal C-arm movement
US11911112B2 (en) 2020-10-27 2024-02-27 Globus Medical, Inc. Robotic navigational system
US11717350B2 (en) 2020-11-24 2023-08-08 Globus Medical Inc. Methods for robotic assistance and navigation in spinal surgery and related systems
US11857273B2 (en) 2021-07-06 2024-01-02 Globus Medical, Inc. Ultrasonic robotic surgical navigation
US11439444B1 (en) 2021-07-22 2022-09-13 Globus Medical, Inc. Screw tower and rod reduction tool
US11911115B2 (en) 2021-12-20 2024-02-27 Globus Medical Inc. Flat panel registration fixture and method of using same

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5389101A (en) * 1992-04-21 1995-02-14 University Of Utah Apparatus and method for photogrammetric surgical localization
US5638819A (en) * 1995-08-29 1997-06-17 Manwaring; Kim H. Method and apparatus for guiding an instrument to a target
US6643535B2 (en) * 1999-05-26 2003-11-04 Endocare, Inc. System for providing computer guided ablation of tissue

Family Cites Families (92)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US599837A (en) * 1898-03-01 Half to frederick pertwee
US4583538A (en) * 1984-05-04 1986-04-22 Onik Gary M Method and apparatus for stereotaxic placement of probes in the body utilizing CT scanner localization
US4991579A (en) * 1987-11-10 1991-02-12 Allen George S Method and apparatus for providing related images over time of a portion of the anatomy using fiducial implants
US5086401A (en) * 1990-05-11 1992-02-04 International Business Machines Corporation Image-directed robotic system for precise robotic surgery including redundant consistency checking
DE69133603D1 (en) * 1990-10-19 2008-10-02 Univ St Louis System for localizing a surgical probe relative to the head
US5603318A (en) * 1992-04-21 1997-02-18 University Of Utah Research Foundation Apparatus and method for photogrammetric surgical localization
US5493574A (en) * 1992-09-24 1996-02-20 Zilog, Inc. Power efficient RAM disk and a method of emulating a rotating memory disk
US5732703A (en) * 1992-11-30 1998-03-31 The Cleveland Clinic Foundation Stereotaxy wand and tool guide
US5517990A (en) * 1992-11-30 1996-05-21 The Cleveland Clinic Foundation Stereotaxy wand and tool guide
DE69432834T2 (en) * 1993-04-26 2004-05-13 St. Louis University Display of the location of a surgical probe
US5803089A (en) * 1994-09-15 1998-09-08 Visualization Technology, Inc. Position tracking and imaging system for use in medical applications
US5682886A (en) * 1995-12-26 1997-11-04 Musculographics Inc Computer-assisted surgical system
US5799055A (en) * 1996-05-15 1998-08-25 Northwestern University Apparatus and method for planning a stereotactic surgical procedure using coordinated fluoroscopy
US6009212A (en) * 1996-07-10 1999-12-28 Washington University Method and apparatus for image registration
US7058822B2 (en) * 2000-03-30 2006-06-06 Finjan Software, Ltd. Malicious mobile code runtime monitoring system and methods
DE19703556A1 (en) * 1997-01-31 1998-08-06 Philips Patentverwaltung Method and arrangement for determining the position in X-ray imaging
US6205411B1 (en) * 1997-02-21 2001-03-20 Carnegie Mellon University Computer-assisted surgery planner and intra-operative guidance system
USD422706S (en) * 1997-04-30 2000-04-11 Surgical Navigation Technologies Biopsy guide tube
US6490474B1 (en) * 1997-08-01 2002-12-03 Cardiac Pathways Corporation System and method for electrode localization using ultrasound
US6226548B1 (en) * 1997-09-24 2001-05-01 Surgical Navigation Technologies, Inc. Percutaneous registration apparatus and method for use in computer-assisted surgical navigation
USD420132S (en) * 1997-11-03 2000-02-01 Surgical Navigation Technologies Drill guide
US6212419B1 (en) * 1997-11-12 2001-04-03 Walter M. Blume Method and apparatus using shaped field of repositionable magnet to guide implant
US6021343A (en) * 1997-11-20 2000-02-01 Surgical Navigation Technologies Image guided awl/tap/screwdriver
DE19829230A1 (en) * 1998-06-30 2000-03-23 Brainlab Med Computersyst Gmbh Method for detecting external contour of objects being treated; prepares flat image for section of area being treated while specifying external contour spots in image surface of section
DE19848765C2 (en) * 1998-10-22 2000-12-21 Brainlab Med Computersyst Gmbh Position verification in camera images
US6285902B1 (en) * 1999-02-10 2001-09-04 Surgical Insights, Inc. Computer assisted targeting device for use in orthopaedic surgery
NZ513919A (en) * 1999-03-17 2001-09-28 Synthes Ag Imaging and planning device for ligament graft placement
DE19917867B4 (en) * 1999-04-20 2005-04-21 Brainlab Ag Method and device for image support in the treatment of treatment objectives with integration of X-ray detection and navigation system
US6190395B1 (en) * 1999-04-22 2001-02-20 Surgical Navigation Technologies, Inc. Image guided universal instrument adapter and method for use with computer-assisted image guided surgery
US6674916B1 (en) * 1999-10-18 2004-01-06 Z-Kat, Inc. Interpolation in transform space for multiple rigid object registration
US6187018B1 (en) * 1999-10-27 2001-02-13 Z-Kat, Inc. Auto positioner
US6381485B1 (en) * 1999-10-28 2002-04-30 Surgical Navigation Technologies, Inc. Registration of human anatomy integrated for electromagnetic localization
US8239001B2 (en) * 2003-10-17 2012-08-07 Medtronic Navigation, Inc. Method and apparatus for surgical navigation
US6379302B1 (en) * 1999-10-28 2002-04-30 Surgical Navigation Technologies Inc. Navigation information overlay onto ultrasound imagery
US7366562B2 (en) * 2003-10-17 2008-04-29 Medtronic Navigation, Inc. Method and apparatus for surgical navigation
US7689014B2 (en) * 2000-01-18 2010-03-30 Z-Kat Inc Apparatus and method for measuring anatomical objects using coordinated fluoroscopy
WO2001064124A1 (en) * 2000-03-01 2001-09-07 Surgical Navigation Technologies, Inc. Multiple cannula image guided tool for image guided procedures
US6535756B1 (en) * 2000-04-07 2003-03-18 Surgical Navigation Technologies, Inc. Trajectory storage apparatus and method for surgical navigation system
US6856827B2 (en) * 2000-04-28 2005-02-15 Ge Medical Systems Global Technology Company, Llc Fluoroscopic tracking and visualization system
US6856826B2 (en) * 2000-04-28 2005-02-15 Ge Medical Systems Global Technology Company, Llc Fluoroscopic tracking and visualization system
DE50000359D1 (en) * 2000-05-09 2002-09-12 Brainlab Ag Method for registering a patient data record from an imaging method in navigation-assisted surgical interventions by means of X-ray image assignment
AU2001294718A1 (en) * 2000-09-25 2002-05-06 Z-Kat, Inc Fluoroscopic registration artifact with optical and/or magnetic markers
EP1190676B1 (en) * 2000-09-26 2003-08-13 BrainLAB AG Device for determining the position of a cutting guide
US6556857B1 (en) * 2000-10-24 2003-04-29 Sdgi Holdings, Inc. Rotation locking driver for image guided instruments
US6718194B2 (en) * 2000-11-17 2004-04-06 Ge Medical Systems Global Technology Company, Llc Computer assisted intramedullary rod surgery system with enhanced features
US6990220B2 (en) * 2001-06-14 2006-01-24 Igo Technologies Inc. Apparatuses and methods for surgical navigation
US20030069581A1 (en) * 2001-10-04 2003-04-10 Stinson David T. Universal intramedullary nails, systems and methods of use thereof
DE60109219T2 (en) * 2001-11-27 2006-02-16 Freni Brembo S.P.A., Curno DEVICE FOR INTERNAL BODY ADJUSTMENT FOR DUOSER MOTORMOTH BRAKES
DE10202091B4 (en) * 2002-01-21 2005-09-08 Siemens Ag Device for determining a coordinate transformation
US8721655B2 (en) * 2002-04-10 2014-05-13 Stereotaxis, Inc. Efficient closed loop feedback navigation
AU2003224997A1 (en) * 2002-04-16 2003-11-03 Michael Conditt Computer-based training methods for surgical procedures
US7065393B2 (en) * 2002-07-11 2006-06-20 Cedara Software Corp. Apparatus, system and method of calibrating medical imaging systems
AU2003257339A1 (en) * 2002-08-26 2004-03-11 Orthosoft Inc. Computer aided surgery system and method for placing multiple implants
US6856828B2 (en) * 2002-10-04 2005-02-15 Orthosoft Inc. CAS bone reference and less invasive installation method thereof
AU2003273680A1 (en) * 2002-10-04 2004-04-23 Orthosoft Inc. Computer-assisted hip replacement surgery
US8052695B2 (en) * 2002-10-11 2011-11-08 Ge Medical Systems Global Technology Company Llc Adjustable instruments for use with an electromagnetic localizer
US7505809B2 (en) * 2003-01-13 2009-03-17 Mediguide Ltd. Method and system for registering a first image with a second image relative to the body of a patient
US7008430B2 (en) * 2003-01-31 2006-03-07 Howmedica Osteonics Corp. Adjustable reamer with tip tracker linkage
EP1615577A2 (en) * 2003-02-04 2006-01-18 Orthosoft, Inc. Cas modular bone reference assembly and limb position measurement system
EP1627272B2 (en) * 2003-02-04 2017-03-08 Mako Surgical Corp. Interactive computer-assisted surgery system and method
US7309339B2 (en) * 2003-02-04 2007-12-18 Howmedica Osteonics Corp. Apparatus for aligning an instrument during a surgical procedure
US6988009B2 (en) * 2003-02-04 2006-01-17 Zimmer Technology, Inc. Implant registration device for surgical navigation system
JP2007512854A (en) * 2003-04-28 2007-05-24 ブラッコ イメージング ソチエタ ペル アチオニ Surgical navigation system (camera probe)
US20050021037A1 (en) * 2003-05-29 2005-01-27 Mccombs Daniel L. Image-guided navigated precision reamers
US20050033117A1 (en) * 2003-06-02 2005-02-10 Olympus Corporation Object observation system and method of controlling object observation system
US8057482B2 (en) * 2003-06-09 2011-11-15 OrthAlign, Inc. Surgical orientation device and method
US20050020909A1 (en) * 2003-07-10 2005-01-27 Moctezuma De La Barrera Jose Luis Display device for surgery and method for using the same
JP4329431B2 (en) * 2003-07-14 2009-09-09 株式会社日立製作所 Position measuring device
US7427272B2 (en) * 2003-07-15 2008-09-23 Orthosoft Inc. Method for locating the mechanical axis of a femur
US7873403B2 (en) * 2003-07-15 2011-01-18 Brainlab Ag Method and device for determining a three-dimensional form of a body from two-dimensional projection images
EP1653856A1 (en) * 2003-08-07 2006-05-10 Xoran Technologies, Inc. Intraoperative imaging system
US7398116B2 (en) * 2003-08-11 2008-07-08 Veran Medical Technologies, Inc. Methods, apparatuses, and systems useful in conducting image guided interventions
US8484001B2 (en) * 2003-08-26 2013-07-09 Voyant Health Ltd. Pre-operative medical planning system and method for use thereof
US20050049485A1 (en) * 2003-08-27 2005-03-03 Harmon Kim R. Multiple configuration array for a surgical navigation system
US7313430B2 (en) * 2003-08-28 2007-12-25 Medtronic Navigation, Inc. Method and apparatus for performing stereotactic surgery
US7756567B2 (en) * 2003-08-29 2010-07-13 Accuray Incorporated Image guided radiosurgery method and apparatus using registration of 2D radiographic images with digitally reconstructed radiographs of 3D scan data
US8571639B2 (en) * 2003-09-05 2013-10-29 Varian Medical Systems, Inc. Systems and methods for gating medical procedures
US7862570B2 (en) * 2003-10-03 2011-01-04 Smith & Nephew, Inc. Surgical positioners
US7275547B2 (en) * 2003-10-08 2007-10-02 Boston Scientific Scimed, Inc. Method and system for determining the location of a medical probe using a reference transducer array
US7835778B2 (en) * 2003-10-16 2010-11-16 Medtronic Navigation, Inc. Method and apparatus for surgical navigation of a multiple piece construct for implantation
US20050085717A1 (en) * 2003-10-21 2005-04-21 Ramin Shahidi Systems and methods for intraoperative targetting
US20050085718A1 (en) * 2003-10-21 2005-04-21 Ramin Shahidi Systems and methods for intraoperative targetting
US20060058615A1 (en) * 2003-11-14 2006-03-16 Southern Illinois University Method and system for facilitating surgery
US20060036162A1 (en) * 2004-02-02 2006-02-16 Ramin Shahidi Method and apparatus for guiding a medical instrument to a subsurface target site in a patient
US20060052691A1 (en) * 2004-03-05 2006-03-09 Hall Maleata Y Adjustable navigated tracking element mount
US20060041181A1 (en) * 2004-06-04 2006-02-23 Viswanathan Raju R User interface for remote control of medical devices
US20060004284A1 (en) * 2004-06-30 2006-01-05 Frank Grunschlager Method and system for generating three-dimensional model of part of a body from fluoroscopy image data and specific landmarks
US7776055B2 (en) * 2004-07-19 2010-08-17 General Electric Company System and method for tracking progress of insertion of a rod in a bone
US7749223B2 (en) * 2004-08-09 2010-07-06 Howmedica Osteonics Corp. Navigated femoral axis finder
US7702379B2 (en) * 2004-08-25 2010-04-20 General Electric Company System and method for hybrid tracking in surgical navigation
US7555331B2 (en) * 2004-08-26 2009-06-30 Stereotaxis, Inc. Method for surgical navigation utilizing scale-invariant registration between a navigation system and a localization system
US8290570B2 (en) * 2004-09-10 2012-10-16 Stryker Leibinger Gmbh & Co., Kg System for ad hoc tracking of an object

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5389101A (en) * 1992-04-21 1995-02-14 University Of Utah Apparatus and method for photogrammetric surgical localization
US5638819A (en) * 1995-08-29 1997-06-17 Manwaring; Kim H. Method and apparatus for guiding an instrument to a target
US6643535B2 (en) * 1999-05-26 2003-11-04 Endocare, Inc. System for providing computer guided ablation of tissue

Cited By (40)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9775681B2 (en) 2002-03-06 2017-10-03 Mako Surgical Corp. Haptic guidance system and method
US10058392B2 (en) 2002-03-06 2018-08-28 Mako Surgical Corp. Neural monitor-based dynamic boundaries
US8010180B2 (en) 2002-03-06 2011-08-30 Mako Surgical Corp. Haptic guidance system and method
US11076918B2 (en) 2002-03-06 2021-08-03 Mako Surgical Corp. Robotically-assisted constraint mechanism
US8571628B2 (en) 2002-03-06 2013-10-29 Mako Surgical Corp. Apparatus and method for haptic rendering
US9775682B2 (en) 2002-03-06 2017-10-03 Mako Surgical Corp. Teleoperation system with visual indicator and method of use during surgical procedures
US8911499B2 (en) 2002-03-06 2014-12-16 Mako Surgical Corp. Haptic guidance method
US9002426B2 (en) 2002-03-06 2015-04-07 Mako Surgical Corp. Haptic guidance system and method
US10231790B2 (en) 2002-03-06 2019-03-19 Mako Surgical Corp. Haptic guidance system and method
US11202676B2 (en) 2002-03-06 2021-12-21 Mako Surgical Corp. Neural monitor-based dynamic haptics
US11298190B2 (en) 2002-03-06 2022-04-12 Mako Surgical Corp. Robotically-assisted constraint mechanism
US9636185B2 (en) 2002-03-06 2017-05-02 Mako Surgical Corp. System and method for performing surgical procedure using drill guide and robotic device operable in multiple modes
US10610301B2 (en) 2002-03-06 2020-04-07 Mako Surgical Corp. System and method for using a haptic device as an input device
US11426245B2 (en) 2002-03-06 2022-08-30 Mako Surgical Corp. Surgical guidance system and method with acoustic feedback
US11298191B2 (en) 2002-03-06 2022-04-12 Mako Surgical Corp. Robotically-assisted surgical guide
US9801686B2 (en) 2003-03-06 2017-10-31 Mako Surgical Corp. Neural monitor-based dynamic haptics
US7840256B2 (en) 2005-06-27 2010-11-23 Biomet Manufacturing Corporation Image guided tracking array and method
US11712308B2 (en) 2006-05-19 2023-08-01 Mako Surgical Corp. Surgical system with base tracking
US10952796B2 (en) 2006-05-19 2021-03-23 Mako Surgical Corp. System and method for verifying calibration of a surgical device
US10028789B2 (en) 2006-05-19 2018-07-24 Mako Surgical Corp. Method and apparatus for controlling a haptic device
US11123143B2 (en) 2006-05-19 2021-09-21 Mako Surgical Corp. Method and apparatus for controlling a haptic device
US11291506B2 (en) 2006-05-19 2022-04-05 Mako Surgical Corp. Method and apparatus for controlling a haptic device
US10350012B2 (en) 2006-05-19 2019-07-16 MAKO Surgiccal Corp. Method and apparatus for controlling a haptic device
US9492237B2 (en) 2006-05-19 2016-11-15 Mako Surgical Corp. Method and apparatus for controlling a haptic device
US11771504B2 (en) 2006-05-19 2023-10-03 Mako Surgical Corp. Surgical system with base and arm tracking
US11844577B2 (en) 2006-05-19 2023-12-19 Mako Surgical Corp. System and method for verifying calibration of a surgical system
US8287522B2 (en) 2006-05-19 2012-10-16 Mako Surgical Corp. Method and apparatus for controlling a haptic device
US9724165B2 (en) 2006-05-19 2017-08-08 Mako Surgical Corp. System and method for verifying calibration of a surgical device
US10070903B2 (en) 2008-01-09 2018-09-11 Stryker European Holdings I, Llc Stereotactic computer assisted surgery method and system
US10105168B2 (en) 2008-01-09 2018-10-23 Stryker European Holdings I, Llc Stereotactic computer assisted surgery based on three-dimensional visualization
EP2191783A1 (en) * 2008-01-09 2010-06-02 Stryker Leibinger GmbH & Co. KG Stereotactic computer assisted surgery based on three-dimensional visualization
US11642155B2 (en) 2008-01-09 2023-05-09 Stryker European Operations Holdings Llc Stereotactic computer assisted surgery method and system
US10588647B2 (en) 2010-03-01 2020-03-17 Stryker European Holdings I, Llc Computer assisted surgery system
US9517107B2 (en) 2010-07-16 2016-12-13 Stryker European Holdings I, Llc Surgical targeting system and method
US10499961B2 (en) 2012-05-23 2019-12-10 Stryker European Holdings I, Llc Entry portal navigation
US9855104B2 (en) 2012-05-23 2018-01-02 Stryker European Holdings I, Llc Locking screw length measurement
US9508149B2 (en) 2012-05-23 2016-11-29 Stryker European Holdings I, Llc Virtual 3D overlay as reduction aid for complex fractures
WO2013174401A1 (en) * 2012-05-23 2013-11-28 Stryker Trauma Gmbh Entry portal navigation
US10039606B2 (en) 2012-09-27 2018-08-07 Stryker European Holdings I, Llc Rotational position determination
EP3241166A4 (en) * 2014-12-31 2018-10-03 Vector Medical, LLC Process and apparatus for managing medical device selection and implantation

Also Published As

Publication number Publication date
US20060173293A1 (en) 2006-08-03
US20060241416A1 (en) 2006-10-26
WO2004069040A3 (en) 2005-03-24

Similar Documents

Publication Publication Date Title
US20060241416A1 (en) Method and apparatus for computer assistance with intramedullary nail procedure
EP1627272B2 (en) Interactive computer-assisted surgery system and method
US20050267353A1 (en) Computer-assisted knee replacement apparatus and method
US20050281465A1 (en) Method and apparatus for computer assistance with total hip replacement procedure
US20070038223A1 (en) Computer-assisted knee replacement apparatus and method
EP1697874B1 (en) Computer-assisted knee replacement apparatus
US20070016008A1 (en) Selective gesturing input to a surgical navigation system
US20170312035A1 (en) Surgical system having assisted navigation
US8706185B2 (en) Method and apparatus for surgical navigation of a multiple piece construct for implantation
US20060200025A1 (en) Systems, methods, and apparatus for automatic software flow using instrument detection during computer-aided surgery
US20070073136A1 (en) Bone milling with image guided surgery
US20070073133A1 (en) Virtual mouse for use in surgical navigation
US20050267722A1 (en) Computer-assisted external fixation apparatus and method
US20070233156A1 (en) Surgical instrument
US20050197569A1 (en) Methods, systems, and apparatuses for providing patient-mounted surgical navigational sensors
US20050267354A1 (en) System and method for providing computer assistance with spinal fixation procedures
WO2004070581A2 (en) System and method for providing computer assistance with spinal fixation procedures
WO2004069041A9 (en) Method and apparatus for computer assistance with total hip replacement procedure
Oentoro A system for computer-assisted surgery with intraoperative ct imaging
Santos-Munné et al. Fluorotactic Surgery using Coordinated Fluoroscopy

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): BW GH GM KE LS MW MZ SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
122 Ep: pct application non-entry in european phase
WWE Wipo information: entry into national phase

Ref document number: 2004708134

Country of ref document: EP

WWW Wipo information: withdrawn in national office

Ref document number: 2004708134

Country of ref document: EP