[go: up one dir, main page]

WO2024214057A1 - System and method for illustrating a pose of an object - Google Patents

System and method for illustrating a pose of an object Download PDF

Info

Publication number
WO2024214057A1
WO2024214057A1 PCT/IB2024/053600 IB2024053600W WO2024214057A1 WO 2024214057 A1 WO2024214057 A1 WO 2024214057A1 IB 2024053600 W IB2024053600 W IB 2024053600W WO 2024214057 A1 WO2024214057 A1 WO 2024214057A1
Authority
WO
WIPO (PCT)
Prior art keywords
patient
image
avatar
instrument
relative
Prior art date
Application number
PCT/IB2024/053600
Other languages
French (fr)
Inventor
Michael R. Leners
Elliot C. SCHMIDT
Brett D. JACKSON
Original Assignee
Medtronic Navigation, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US18/630,474 external-priority patent/US20240341860A1/en
Application filed by Medtronic Navigation, Inc. filed Critical Medtronic Navigation, Inc.
Publication of WO2024214057A1 publication Critical patent/WO2024214057A1/en

Links

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/20Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/25User interfaces for surgical systems
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B17/00Surgical instruments, devices or methods
    • A61B17/00234Surgical instruments, devices or methods for minimally invasive surgery
    • A61B2017/00238Type of minimally invasive operation
    • A61B2017/00243Type of minimally invasive operation cardiac
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/20Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
    • A61B2034/2046Tracking techniques
    • A61B2034/2051Electromagnetic tracking systems
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/20Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
    • A61B2034/2046Tracking techniques
    • A61B2034/2055Optical tracking systems
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • A61B90/37Surgical systems with images on a monitor during operation
    • A61B2090/374NMR or MRI
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • A61B90/37Surgical systems with images on a monitor during operation
    • A61B2090/376Surgical systems with images on a monitor during operation using X-rays, e.g. fluoroscopy
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • A61B90/37Surgical systems with images on a monitor during operation
    • A61B2090/376Surgical systems with images on a monitor during operation using X-rays, e.g. fluoroscopy
    • A61B2090/3762Surgical systems with images on a monitor during operation using X-rays, e.g. fluoroscopy using computed tomography systems [CT]
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • A61B90/37Surgical systems with images on a monitor during operation
    • A61B2090/378Surgical systems with images on a monitor during operation using ultrasound
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/08Clinical applications
    • A61B8/0883Clinical applications for diagnosis of the heart
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/42Details of probe positioning or probe attachment to the patient
    • A61B8/4245Details of probe positioning or probe attachment to the patient involving determining the position of the probe, e.g. with respect to an external reference frame or to the patient
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/46Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient
    • A61B8/461Displaying means of special interest
    • A61B8/463Displaying means of special interest characterised by displaying multiple images or images and diagnostic data on one display

Definitions

  • the present disclosure relates to acquisition of image data of a subject, and particularly to acquisition and display of image data collected from a tracked imaging system.
  • An imaging system can be used to image various portions of a subject.
  • the subject can include a patient, such as a human patient.
  • the portions selected to be imaged can be internal portions that are covered by skin or other tissue.
  • a location of imaged portions of the subject may be selected to be known within the imaging system. The locations can be defined or established relative to instruments placed in the subject (e.g., a location of a heart wall relative to a catheter) or a location of the imaged portion relative to the instrument acquiring the image data.
  • An imaging modality may be used to generate image data that are used to render images of an interior of the subject.
  • x-rays may be used to generate image data of an interior of a subject.
  • the images may be displayed for viewing by a user.
  • the subject disclosure relates to illustrating a position, including a pose of an instrument relative to a patient.
  • the patient may be a human or nonhuman patient. Further, the patient or subject may be a non-living object where the location of an instrument within the subject is selected to be tracked.
  • a representation of the patient may be fit or morphed to illustrate or represent a size of the patient. Therefore, relative portions of the patient, such as relative portions of the anatomy, are determined or known relative to one another based upon the patient that is the current patient during a procedure. An avatar or indicia of the patient may be generated or altered to a specific patient substantially in real time during a procedure.
  • An instrument can be tracked relative to the patient.
  • a representation of the instrument may be superimposed on the avatar of the patient.
  • the avatar of the patient may provide a visual illustration of various portions of the anatomy of the patient without requiring image data being acquired of the patient.
  • the display being viewed by the user may be generated without requiring image data of the patient, such as fluoroscopy image data.
  • FIG. 1 is an environmental view of an exemplary imaging and navigation system
  • Fig. 2A is a screen shot of a representation of an instrument at a tracked pose relative to an avatar of a subject, the screen shot including an image taken at a first orientation;
  • Fig. 2B is similar to Fig. 2A, but with the image at a second orientation that is about 90° relative to the first orientation;
  • FIG. 3 is a schematic view of a patient in an operating theater
  • FIG. 4 is a schematic view of a user and a patient in an operating theater, according to various embodiments
  • Fig. 5 is a detailed view of an instrument being moved relative to a patient;
  • Fig. 6 is a screen shot of a display illustrating a position of an instrument relative to an avatar;
  • FIG. 7 is a flowchart of a method, according to various embodiments.
  • Fig. 8 is a schematic illustration of an atlas and an image
  • FIG. 9 is an illustration of images in two planes of a heart, according to various embodiments including ultrasound;
  • FIGs. 10A and 10B illustrate movement of a tracked instrument over time
  • FIG. 11 is a flowchart illustrating a method of showing movement of a tracked instrument over time, according to various embodiments
  • Fig. 12A illustrates a tracked pose of an instrument relative to an image in a plane in two orientations, according to various embodiments
  • Fig. 12B illustrates a tracked pose of an instrument relative to an image in a plane in two orientations, according to various embodiments
  • FIG. 13 illustrates an exemplary screen shot with right and left panels, on the right is an avatar of a subject with a housing of an imaging device and a field of view of the imaging device overlaid on the avatar, on the left is an image captured by the imaging device;
  • Fig. 14 illustrates an exemplary screen shot with right, left, and center panels
  • the center panel includes an avatar of a subject with a housing of an imaging device overlaid on the avatar, a first field of view in the center panel corresponds to an image captured by the imaging device shown in the left panel and a second field of view in the center panel corresponds to an image captured by the imaging device shown in the right panel;
  • FIG. 15 illustrates an exemplary screen shot with four panels, each panel shows a different orientation of a heart icon representing a heart being scanned with an imaging device, areas of the heart that have been scanned are displayed with different shading, coloring, etc. to distinguish from areas of the heart that have not yet been scanned;
  • Fig. 16 illustrates an exemplary screen shot with a sensor icon and a range icon overlaid on the avatar to identify a position and range of a sensor
  • Fig. 17 illustrates an exemplary screen shot with a left panel including a first respiration sensor icon and a second respiration sensor icon arranged on the avatar, and a right panel including a graphical representation of movement between first and second respiration sensors on a subject’s body.
  • Fig. 1 is a diagram illustrating an overview of a navigation system 10 that can be used for various procedures.
  • the navigation system 10 can be used to track the location of an item, such as an implant or an instrument, and at least one imaging system 12 relative to a subject, such as a subject 14 that may include a human patient and/or other living or non-living subject.
  • the navigation system 10 may be used to navigate any type of instrument, implant, or delivery system, including: catheters, stylets, leads for cardiac rhythm management devices such as pacemakers, leadless pacemakers and delivery systems therefor, guide wires, arthroscopic systems, ablation instruments, stents, orthopedic implants, spinal implants, deep brain stimulation (DBS) probes, mechanical parts, Transesophageal Echocardiography (TEE), intra-cardiac echocardiography (ICE), etc.
  • Non-human or non-surgical procedures may also use the navigation system 10 to track a non-surgical or non-human intervention of the instrument or imaging device.
  • the instruments may be used to navigate or map any region of the body.
  • the navigation system 10 and the various tracked items may be used in any appropriate procedure, such as one that is generally minimally invasive or an open procedure.
  • the navigation system 10 can interface with or integrally include an imaging system 12 that is used to acquire pre-operative, intra-operative, or postoperative, or real-time image data of the patient 14.
  • the imaging system 12 can be an ultrasound imaging system (as discussed further herein) that has a tracking device 22 attached thereto.
  • the imaging system 12 may also include any suitable CT system, x-ray system, C-arm, O-arm® imaging system, magnetic resonance imaging system, etc.
  • the tracking device 22 may be tracked with the tracking system to determine a pose of the imaging system 12.
  • the pose may include an orientation (e.g., three or more degrees of orientation (e.g., yaw, pitch, and roll) and/or a position (e.g., three degrees of freedom in physical space (e.g., x-axis, y-axis, and z-axis).
  • the tracking system 10 may further determine appropriate pose information regarding the tracking device 22.
  • the pose of the imaging system 12 can then be determined based on the tracked pose of the tracking device 22.
  • the imaging system 12 may be used to generate image data to provide images for viewing with a selected display device 26, which may any appropriate display device including an augmented and/or virtual reality display device worn/used by the user 18 such as those disclosed in U.S. Pat. App. Pub. No. US2018/0078316A1 , incorporated herein by reference.
  • the navigation system 10 can be used to track various tracking devices, as discussed herein, to determine locations of the patient 14.
  • the tracked poses of the patient 14 can be used to determine or select images for display to be used with the navigation system 10.
  • the initial discussion, however, is directed to the navigation system 10 and the exemplary imaging system 12.
  • the imaging system 12 includes an ultrasound (US) imaging system with an US housing 16 that is held by a user 18 while collecting image data of the subject 14. It will be understood, however, that the US housing 16 can also be held by a stand or robotic system while collecting image data.
  • the US housing 16 and included transducer can be any appropriate US imaging system 12, such as the M-TURBO® sold by SonoSite, Inc. having a place of business at Bothell, Washington.
  • the tracking device 22 may be any appropriate tracking device such as an electromagnetic (EM) tracking device and/or an optical tracking device.
  • the tracking devices can be used together (e.g., to provide redundant tracking information) or separately. Also, only one of the two tracking devices may be present. It will also be understood that various other tracking devices can be associated with the US housing 16, as discussed herein, including acoustic, ultrasound, radar, electrical impedance, and other tracking devices. Also, the tracking device 22 can include linkages or a robotic portion that can determine a location relative to a reference frame. Any suitable tracking device of any suitable tracking system may be used, such as, tracking devices and systems as understood by one skilled in the art.
  • a secondary imaging system may also be presented and/or used to generated image data of the patient 14.
  • the second imaging system may include an O-arm® imaging device sold by Medtronic Navigation, Inc. having a place of business in Louisville, Colorado, USA, a C-arm imaging system, or other appropriate imaging system.
  • the second imaging system can include those disclosed in U.S. Pat. Nos. 7,188,998; 7,108,421 ; 7,106,825; 7,001 ,045; and 6,940,941 ; all of which are incorporated herein by reference.
  • the second imaging system need not be present and the imaging system 12, which may be only the US transducer, may be the only imaging system to generate images to be displayed.
  • various other image data may also be displayed such as an avatar 120 of the patient 14.
  • the imaging system 12 may be tracked, as discussed above.
  • Various tracking systems may include and/or require calibration of the imaging system.
  • the pose of the tracking device 22 relative to a plane of the US imaging system may be determined and/or known.
  • Various systems and methods are disclosed in U.S. Patent Nos. 6,379,302; 6,669,635; 6,968,224; 7,085,400; 7,831 ,082; 8,320,653; 8,811 ,662; and 9,138,204 all of which are incorporated herein by reference.
  • the patient 14 can be fixed in a selected pose relative to a selected object, such as onto an operating table 40, but is not required to be fixed to the table 40.
  • the table 40 can include a plurality of straps 42 or other fixation devices.
  • the straps 42 can be secured around the patient 14 to fix the patient 14 relative to the table 40.
  • Various apparatuses may be used to position the patient 40 in a static position on the operating table 40. Examples of such patient positioning devices are set forth in commonly assigned U.S. Pat. App. No. 10/405,068, published as U.S. Pat. App. Pub. No. 2004/0199072 on October 7, 2004, entitled “An Integrated Electromagnetic Navigation And Patient Positioning Device”, filed April 1 , 2003 which is hereby incorporated by reference.
  • Other known apparatuses may include a Mayfield® clamp.
  • the navigation system 10 includes at least one tracking system.
  • the tracking system can include at least one localizer.
  • the tracking system can include an electromagnetic (EM) localizer 50.
  • the tracking system can be used to track instruments relative to the patient 14 or within a navigation space.
  • the navigation system 10 can use image data from the imaging system 12 and information from the tracking system to illustrate locations of the tracked instruments, as discussed herein.
  • the tracking system can also include a plurality of types of tracking systems including an optical localizer 52 in addition to and/or in place of the EM localizer 50.
  • the EM localizer 50 can communicate with or through a localizer communication 54 that may be wired or wireless.
  • the EM localizer 50 may also include and/or take the form of an alternative pad or flat EM localizer 55.
  • the optical localizer 52 and the EM localizer 50 can be used together to track multiple instruments or used together to redundantly track the same instrument.
  • Various tracking devices can be tracked and the information can be used by the navigation system 10 to allow for an output system to output, such as a display device to display, a position of an item.
  • tracking devices can include a patient or reference tracking device 56, also known as a dynamic reference frame, to track the patient 14, an instrument tracking device 60 to track an instrument 62, and/or other appropriate tracking devices for one or more portions.
  • Patient or reference tracking device 56 and instrument tracking device 60 may include those disclosed in U.S. Pat. Nos. 8,060,185 and 8,644,907, both incorporated herein by refence.
  • the tracking devices allow selected portions of the operating theater to be tracked relative to one another with the appropriate tracking system, including the optical localizer 52 and/or the EM localizer 50.
  • the reference tracking device 56 can also or alternatively be positioned on an instrument and positioned within the patient 14, such as within a heart 15 of the patient 14.
  • any of the tracking devices 22, 56, 60 can be optical or EM tracking devices, or both, depending upon the tracking localizer used to track the respective tracking devices. It will be further understood that any appropriate tracking system can be used with the navigation system 10.
  • Alternative tracking systems can include radar tracking systems, acoustic tracking systems, ultrasound tracking systems, electrical impedance tracking systems, radio frequency beacon, and the like. Exemplary tracking systems include those disclosed in U.S. Pat. Nos. 7,676,268; 8,532,734; and 8,494,608, all incorporated herein by reference.
  • Each of the different tracking systems can be respective different tracking devices and localizers operable with the respective tracking modalities. Also, the different tracking modalities can be used simultaneously as long as they do not interfere with each other (e.g., an opaque member blocks a camera view of the optical localizer 52).
  • An exemplary EM tracking system can include the STEALTHSTATION® AXIEMTM Navigation System, sold by Medtronic Navigation, Inc. having a place of business in Louisville, Colorado. Exemplary tracking systems are also disclosed in U.S. Patent No. 7,751 ,865, issued July 6, 2010 and entitled “METHOD AND APPARATUS FOR SURGICAL NAVIGATION"; U.S. Patent No. 5,913,820, titled “Position Location System,” issued June 22, 1999 and U.S. Patent No. 5,592,939, titled “Method and System for Navigating a Catheter Probe,” issued January 14, 1997, all herein incorporated by reference.
  • shielding or distortion compensation systems can include those in U.S. Pat. No. 7,797,032, issued on September 14, 2010 and U.S. Pat. No. 6,747,539, issued on June 8, 2004; distortion compensation systems can include those disclosed in U.S. Pat. No. 10/649,214, filed on January 9, 2004, published as U.S. Pat. App. Pub. No. 2004/01 16803, all of which are incorporated herein by reference.
  • the EM localizer 50 and the various tracking devices can communicate through an EM controller.
  • the EM controller can include various amplifiers, filters, electrical isolation, and other systems.
  • the EM controller can also control the coils of the EM localizer 50 to either emit or receive an EM field for tracking.
  • a wireless communications channel such as that disclosed in U.S. Patent No. 6,474,341 , entitled “Surgical Communication Power System,” issued November 5, 2002, herein incorporated by reference, can be used as opposed to being coupled directly to the EM controller.
  • the EM controller can be incorporated into a navigation processing system 70.
  • the tracking system may also be or include any appropriate tracking system, including a STEALTHSTATION® TRIA®, TREON®, and/or S7TM Navigation System having an optical localizer, similar to the optical localizer 52, sold by Medtronic Navigation, Inc. having a place of business in Louisville, Colorado.
  • STEALTHSTATION® TRIA® similar to the optical localizer 52, sold by Medtronic Navigation, Inc. having a place of business in Louisville, Colorado.
  • alternative tracking systems are disclosed in U.S. Patent No. 5,983,126, to Wittkampf et al. titled “Catheter Location System and Method,” issued November 9, 1999, which is hereby incorporated by reference.
  • Other tracking systems include an acoustic, radiation, radar, electrical impedance, etc. tracking or navigation systems.
  • the navigation system 10 can include a navigation processing unit or module 74 that can communicate or include a navigation memory 76, which may be included in the navigation processing system 70.
  • the navigation processing system 70 may further include a display device 77.
  • the navigation processing unit 74 can include a processor and memory (e.g., microprocessor, a central processing unit, etc.). In various embodiments, the navigation processing unit may execute instructions to determine one or more poses of the tracking devices based on signals from the tracking devices.
  • the navigation processing unit 74 can receive information, including image data, from the imaging system 12 and tracking information from the tracking systems, including the respective tracking devices and/or the localizers 50, 54. Image data can be displayed as an image 78 on the display device 26.
  • the display device may be separate from and/or integrated into the navigation system, thus, the display device 26 may include the display device 77.
  • the navigation system 70 can include appropriate input devices, such as a keyboard 84. It will be understood that other appropriate input devices can be included, such as a mouse, a foot pedal 88 or the like which can be used separately or in combination.
  • all of the disclosed processing units or systems can be a single processor module (e.g., a single central processing chip) that can execute different instructions to perform different tasks.
  • An image processing unit or module can process image data from the imaging system 12 and a separate first image processor (not illustrated) can be provided to process or pre-process image data from the imaging system 12. The image data from the image processor can then be transmitted to the navigation processor 74.
  • the imaging systems need not perform any image processing and the image data can be transmitted directly to the navigation processing unit 74.
  • the navigation system 10 may include or operate with a single or multiple processing centers or units that can access single or multiple memory systems based upon system design.
  • the imaging system 12 can generate image data that may be used to compose the image 78 and define an image space that can be registered to a patient space or navigation space that is defined by and/or relative to the patient 14.
  • the position of the patient 14 relative to the imaging system 12 can be determined by the navigation system 10 with the patient tracking device 56 and the imaging system tracking device(s) 22 to assist in and/or maintain registration. Accordingly, the position of the patient 14 relative to the imaging system 12 can be determined.
  • Registration can occur by matching fiducial points in image data with fiducial points on the patient 14. Registration of image space to patient space allows for the generation of a translation map between the patient space and the image space. According to various embodiments, registration can occur by determining points that are substantially identical in the image space and the patient space. The identical points can include anatomical fiducial points or implanted fiducial points. Exemplary registration techniques are disclosed in U.S. Pat. App. No. 12/400,273, filed on March 9, 2009, now published as U.S. Pat. App. Pub. No. 2010/02281 17 and in U.S. Pat. No. 9,737,235, issued August 22, 2017, both incorporated herein by reference.
  • the imaging system 12 can be used with an un-navigated or navigated procedure.
  • a localizer and/or digitizer including either or both of an optical localizer 52 and/or an electromagnetic localizer 50, 55 can be used to generate a field and/or receive and/or send a signal within a navigation domain relative to the patient 14.
  • the navigated space or navigational domain relative to the patient 14 can be registered to the image 78.
  • Correlation is to allow registration of a navigation space defined within the navigational domain and an image space defined by the image 78.
  • the patient tracker or dynamic reference frame 56 can be connected to the patient 14 to allow for a dynamic registration and maintenance of registration of the patient 14 to the image 78.
  • the navigation system 10 with or including the imaging system 12 can be used to perform selected procedures. Selected procedures can use the image data generated or acquired with the imaging system 12. Further, the imaging system 12 can be used to acquire image data at different times relative to a procedure. As discussed herein, image data can be acquired of the patient 14 prior to the procedure for collection of automatically registered image data or cine loop image data. Also, the imaging system 12 can be used to acquire images for confirmation of a portion of the procedure. Thus, image data may be acquired at any appropriate time and may be registered to the patient 14.
  • a graphic representation 90 e.g., an icon, indicium, animation or other or visual representation
  • the image 78 may be an appropriate image and may include one or more 2D images, such as 2D images that are acquired at different planes. Images may also be a 3D image, or any appropriate image as discussed herein.
  • the imaging plane of the US imaging system 12 can also be determined.
  • imaged portions can be located within the patient 14.
  • a position of an imaged portion of the heart 15, or other imaged portion can also be tracked as disclosed in U.S. Patent Nos. 6,379,302; 6,669,635; 6,968,224; 7,085,400; 7,831 ,082; 8,320,653; 8,811 ,662; and 9,138,204 all of which are incorporated herein by reference.
  • the patient 14 may be positioned on the table 40, as noted above.
  • the display 26 and/or the display 77 may display various information regarding the patient 14 and/or other information selected by the user 18.
  • the display 26 may illustrate the image 78 that may be acquired with the imaging system 12.
  • the image 78 may also be displayed on the display 77.
  • the instrument 62 may be tracked and the graphic representation 90 of the instrument 62 may be displayed on the display 26, such as relative to the image 78.
  • the graphic representation 90 may also be displayed relative to a patient avatar 120.
  • the patient avatar 120 may be illustrated on a selected display, as illustrated in Figs. 2A and 2B, such as the display 26.
  • the patient avatar 120 may be based on a general avatar that includes various features of the patient 14, as noted herein, but is sized to the current and specific patient.
  • the avatar 120 may be illustrated as a two-dimensional (2D) image and/or a three-dimensional (3D) image.
  • the avatar 120 may be any appropriate shape to assist in identifying various positions and relative positions of portions of the patient 14.
  • the avatar 120 is sized to the specific patient as disclosed herein.
  • the patient avatar 120 may generally illustrate a portion of the patient 14 and may be sized relative to the patient 14.
  • the patient avatar may illustrate an inferior portion of a neck 122, a pectoral region 124, and other appropriate regions of the patient 14.
  • the patient avatar 120 may be displayed on the display device 26 to illustrate a general representation of the patient 14 without requiring an image acquisition of the same portion of the patient 14. Therefore, the patient avatar 120 may provide a general roadmap or indication of the position of various tracked portions, as discussed further herein.
  • the avatar 120 may however, as discussed herein, be morphed to match a relative size of the patient 14 to provide an indication of relative anatomical portions of the patient 14 and a pose of the tracked portions, such as the instrument 62, relative thereto.
  • the avatar 120 may be generally sized to the patient 14 according to one or more appropriate procedures.
  • the avatar 120 may be sized or morphed in real time to match relative dimensions of the patient 14.
  • the relative positions of the anatomic portions may include a relative position of a suprasternal notch and an exterior shoulder or inferior jugular vein, or any other suitable anatomical location(s).
  • the user 18 may provide or include information regarding the patient 14.
  • the navigation processing system 70 may include various inputs that allow the user 18 to input information. Therefore, as illustrated in Fig. 3, the user 18 may input information such as a height, a weight, and a sex of the patient 14.
  • the input information is used by the processor module 74 to execute selected instructions to size the patient avatar 120 to the actual patient 14.
  • the resizing of the avatar 120 may be based upon recalling of a selected or known predetermined size from a database, reconfiguring or resizing the patient avatar 120 based upon the input information, or other appropriate information. Therefore, the avatar 120 may be sized to substantially match the patient 14 without acquiring an image of the patient 14 for display on the display device 26.
  • the user 18 may also identify or size the avatar 120 to the patient 14 with a tracked indicator or tracked member 130.
  • the tracked indicator 130 which may be an indicator or probe usable to touch a surface of the patient 14, may be moved by the user 18 to various predetermined positions on the patient 14 as indicated on the avatar 120.
  • the tracked indicator 130 may be tracked with any appropriate tracking system, including those discussed herein according to various embodiments such as the localizer system of U.S. Patent No. 5,983,126, hereby incorporated by reference.
  • the user 18 may move the tracked indicator 130 to various points, such as the exterior shoulder portions 132’, 134’ as illustrated by the icons 132 and 134.
  • the user 18 may move the tracked indicator 130 to other portions such as a suprasternal notch 136 and a sternum 140 of the patient 14 or outer boundary positions. It is understood that the tracked indicator 130 may be moved to any appropriate position and the above-noted positions are merely exemplary. Nevertheless, based upon the tracked position of the tracked indicator 130, the avatar 120 may be sized to morph the identified points to a representation of the patient 14 for display on the display device 26. Also, the points on the avatar 120 may be registered to the patient 14 by identifying them in the patient space relative to the patient 14 and identifying them on the avatar 120.
  • the avatar 120 may also be morphed to actual measurements taken of the patient 14. For example, a distance between the shoulder points 132’, 134’ on that the patient 14 may be measured, such as with a tape measure or other measuring device. The measured distance may be input within the navigation processing system at 77. The avatar 120 may be sized based upon the measurements.
  • the avatar 120 may be matched or morphed to substantially mimic the shape of the patient 14.
  • the avatar 120 may then be used as an illustration of a pose of the tracked instrument, such as the instrument 62, during a procedure on the patient 14.
  • the avatar 120 may provide a representation of a position of the tracked instrument 62 without requiring an image acquisition of the patient 14, such as with a fluoroscopic imaging system.
  • the avatar 120 as it is morphed to the specific patient may or will be displayed as a patient specific avatar.
  • the avatar 120 may include general information, but is displayed and registered to the patient 14 as a patient specific avatar 120 including relevant locations and relative positions of various portions of the patient 14.
  • the specific patient avatar illustrates the relative positions of the anatomy of the patient 14 that is the current patient. For example, the position of the heart 15 relative to the suprasternal notch.
  • the avatar 120 may be registered to the patient 14 in any appropriate manner, including those discussed above.
  • the tracked indicator 130 is tracked to determine positions of the patient 14 for marking or changing the avatar 120.
  • the patient 14 after the various points are determined in both the patient 14 and the avatar 120 may be registered thereto.
  • the patient tracker 56 may be positioned on the patient 14 at an appropriate position.
  • the position of the patient tracker 56 may be at the suprasternal notch 136’ which relates to a point on the avatar 120, such as the suprasternal notch 136.
  • the position of the suprasternal notch 136 in the avatar 120 may be determined or known to be the position of the patient tracker 56 to allow for registration of the avatar 120 relative to the patient 14.
  • the patient tracker 56 may maintain registration even when the patient 14 moves relative to the localizer(s). It is understood that any appropriate number of the patient trackers 56 may be utilized, including each may be used for registration and/or more than one of the patient trackers 56 may be used.
  • the localizers may also or alternatively be fixed or positioned relative to the patient 14 in an appropriate manner.
  • the localizer 58 may be fixed relative to the table 14 such that it is immovable during the procedure.
  • the field generated with the EM localizer 58 can be at a known position relative to the patient 14 and allow for a registration of the avatar 120 to the patient 14 based upon the known position of the EM localizer 58 relative to the patient 14.
  • the localizer 55 may also be positioned under the patient 14, such as between the patient 14 and the bed 40. Localizer 55 may include field generating portions, such as coils, that generate a field relative to the patient 14.
  • the flat or panel localizer 55 may be positioned at a known position relative to the patient 14, such as aligning the shoulder edges 132’, 134’ to known or predetermined portions of the panel localizer 55. As the panel localizer 55 is at a known position relative to the patient 14, and the avatar 120 is morphed to the patient and 14, the avatar 120 may be registered to the patient 14 due to the position of the panel localizer 55.
  • the avatar 120 may be registered to the patient 14. Registration of the avatar 120 relative to the patient 14 allows the avatar 120 to be used to assist navigation and guiding of the instrument 62, or any appropriate instrument relative to the patient 14. As the tracked position of the instrument 62 is determined with the navigation system 10, the pose of the instrument 62 may be illustrated with a graphical representation, such as the graphic representation 90 relative to the image 78 and/or the avatar 120.
  • the instrument’s graphic representation 90 is displayed superimposed on the avatar 120.
  • the user 18 may view the display 26 and understand the tracked position of the instrument 62 relative to the patient 14 by viewing the graphic representation 90 of the instrument 62 in the avatar 120 on display 26.
  • the user 18 may then understand the relative position of the instrument 62 to the patient 14 and determine that the instrument 62 is generally moving and/or is positioned in a selected or procedure planned position.
  • the user 18, thus, may understand this by viewing the display 26 and the avatar 120 without requiring additional or any image data being acquired of the patient 14.
  • image data may be acquired of the patient 14.
  • the imaging system 12 may acquire image data of the patient 14.
  • the imaging system 12 may generally be a non-ionizing imaging system, such as an ultrasound imaging system.
  • the ultrasound image may generally be collected at a single plane relative to the housing 16.
  • the US imaging system 12 with the tracker 22 may be tracked relative to the patient 14 so that the image may also be registered to the patient 14.
  • the image 78 may also be displayed superimposed on the avatar 120, such as the image superimposition 78i.
  • the image may also be displayed separately from the avatar, as also illustrated in Figs. 2A and 2B.
  • Fig. 2A illustrates the image 78 at a first orientation.
  • Fig. 2B illustrates the image 78 at a second orientation that is about 90 relative to the first orientation.
  • the operator of the imaging system 12 may toggle between views to select the view that they find most suitable without moving the housing 16.
  • the image 78 may be registered to the patient 14, a single image acquisition may occur and the image 78 may be displayed for the procedure. Further, as the image 78 is registered to the patient the graphic representation 90 of the instrument 62 may be displayed at the tracked pose on the image 78.
  • the US imaging system 12 may be used at one point in time to acquire the image data, but it may be used to illustrate the selected portion of the patient 14, such as the heart, for an entire procedure.
  • a constant imaging of the patient 14 is not required to illustrate or know a position of the instrument 62 as the instrument 62 is tracked and the image 78 is registered to the patient 14 due to the US tracker [0070]
  • the user 18 may perform the procedure on the patient 14 with the tracked instrument(s).
  • the procedure may be any appropriate procedure, such as placing a pacing lead and/or a catheter to assist in placing the lead. Other appropriate procedures may also be performed.
  • the user 18 may form an incision, such as an incision 150 on the patient 14, and move the instrument 62 into the patient 14 through the incision 150.
  • the tracked portion 60 on the instrument 62 may be tracked so that a position of the instrument 62 may be displayed relative to the avatar 120 and/or the image 78, as noted above.
  • the instrument 62 may then be moved relative to the patient 14.
  • the instrument 62 may be a catheter and/or lead for an implant, or a stylet or a delivery system for an implant such as a leadless pacemaker or a stent.
  • an implant, including an electrode may be positioned relative to one or more portions of the heart 15 of the patient 14.
  • the position of the instrument 62 may be illustrated with the graphic representation 90 over time and/or instantaneously relative to the avatar 120.
  • the graphic representation 90 illustrates the pose of the instrument 62 tracked relative to the patient 14.
  • the graphic representation 90 may be viewed by the user 18 to understand a real-time position of the instrument 62 within the patient 14 via the avatar 120. Again, the avatar 120 may be generated without requiring image data of the patient 14, such as with fluoroscopy.
  • the avatar 120 may also assist the user 18 in identifying or determining an orientation of the instrument 62 as being in a selected or in a not- selected pose relative to the patient 14. For example, as illustrated in Fig.
  • the avatar 120 may be viewed by the user 18 during a procedure.
  • the user 18 may view the representation 90 of the instrument 62 as it moves in a selected direction and/or position.
  • the tracked pose of the instrument 62 may be updated in substantially real time and displayed relative to the avatar 120. Therefore, the graphic representation 90 of the instrument 62 when it is moving in a not-selected position or and/or pose may also be illustrated and understood by the user 18.
  • a representation 90x may illustrate movement and/or pose toward a neck 122 of the patient 14. This may be a not-selected pose when the instrument 62 is selected to be moving toward the heart 15 of the patient 14. Therefore, the user 18 may view the graphic representation 90x and understand that the instrument 62 is moving away from the heart 15 of the patient 14.
  • a representation of the heart 15 may also be displayed on the avatar 120 including the image 78 and/or with a heart representation 15i that is included in the avatar 120 and appropriately located with the patient specific information.
  • the user 18 may view and understand the position of the instrument 62 during the placement of an implant, which may include, for example, a stimulation lead for a selected system.
  • the lead may include a pacing electrode be for a cardiac stimulation system, such as the AZURE®, ADAPTA®, or other appropriate heart pacemaker or defibrillation systems sold by Medtronic, Inc. having a place for business in Minnesota, USA.
  • a method 200 for determining and displaying a patient specific avatar is illustrated. The process may begin in Block 210 and include recalling a standard avatar in Block 214.
  • the processor may recall a standard avatar including various anatomical features, including exterior dimensions of a patient's anatomy and selected internal anatomy such as a heart, physical location of the heart, jugular vein, or the like.
  • Information regarding the patient may be received in Block 218.
  • the information regarding the patient may be any appropriate information, including that discussed above. Therefore, the information may include general dimensions of the current patient, including a height, weight, sex, or the like.
  • Information received in Block 218 may also include specific tracked points of predetermined portions (e.g., shoulder edges) of the patient 14. Further the received information may include measurements of the patient. Regardless, the received information of the patient may be patient specific information regarding the patient for a current or real-time procedure.
  • a patient specific avatar is then generated in Block 230.
  • the generation of a patient specific avatar may include morphing or fitting the standard avatar based upon the received information.
  • the method may include altering or scaling dimensions of the standard avatar based upon a ratio given the information received regarding the current patient. The ratio may be based upon the known dimensions of the current patient based upon known or average dimensions of various anatomical portions relative to one another based upon the dimensions of the specific patient. This may also include having predetermined average positions and relative positions of various anatomical features.
  • the tracked position of the instrument may identify specific anatomical locations.
  • the standard avatar may then be morphed based upon the tracked specific locations for the specific patient.
  • the avatar may be morphed based upon selected instructions including an algorithm to generate the patient specific avatar in Block 230.
  • the patient specific avatar may be displayed in Block 234.
  • the display 26 may display the patient specific avatar for selected procedures.
  • the display of the patient specific avatar may allow the user 18 to identify and understand relative positions or locations of various portions of the current patient 14.
  • An instrument may then be tracked, optionally and according to various embodiments, in Block 238.
  • the tracking of the instrument is an optional part of the method 200, such tracking may allow for information regarding a pose of an instrument to be determined.
  • the method 200 may include receiving tracking information in Block 240.
  • Receiving tracking information in Block 240 may include tracking one or more tracking devices to determine a pose of various elements or portions, such as the pose of the instrument 62 via the instrument tracker 60 and/or the patient 14 via the patient tracker 56.
  • the patient specific avatar may be registered to the patient based upon the known location of one or more specific portion(s) of the patient, such as the positioning of the patient tracker 56 in the suprasternal notch of the patient 14.
  • the suprasternal notch of the avatar 120 may then be registered for determination of the pose of the instrument 62 relative to the patient tracker 56. Therefore, the pose of the instrument 62 may be displayed on the display device 26 such as the superimposed graphic representation 90 on the avatar 120.
  • the pose of the instrument may be superimposed on the avatar 120 in Block 244. Again, this allows a user 18 to view and understand a pose of the tracked instrument 62 on a representation of the patient 14 without requiring image data to be acquired to the patient 14. While image data of the patient 14 may be acquired, such as with the US imaging system 12, such image information is not required. The image data information from the US imaging system 12 may assist in finalizing or determining a specific location of the instrument 62 relative to a selected portion, such as a specific portion of the heart 15 of the patient 14. According, only a selected portion or area of the patient may be imaged. However, the position of the instrument 62 may be determined without the image data. Thus, the user may understand a position of the instrument 62, including a movement of the instrument 62 toward a selected position, such as the RV septum of the patient 14.
  • the method 200 may then end in Block 250. And ending the method 250 it is understood that the method 200 may be repeated as often as necessary or selected to allow for up to date and/or real-time tracking and illustration of a pose of the instrument 62 relative to the avatar 120. Thus, the method 200 may be repeated at a selected rate and for a selected period, such as selected by the user 18.
  • moving the instrument 62 relative to the patient 14, such as relative to the heart 15, may be accomplished while tracking a pose of the instrument 62.
  • tracking a pose of the instrument 62 can involve attempting to determine a physical location and/or orientation of the instrument 62 in the physical space.
  • the tracked posed may include a location which includes X, Y, and Z axis positions and orientation which may include yaw, pitch, and roll orientations in space.
  • the physical space may also be referred to as navigation space and is generally defined relative to the patient 14 and/or the heart 15.
  • the patient tracker 56 allows the pose of the patient 14, including the heart 15, to be tracked. Tracking a pose relative to a portion of the patient 14, including the heart 15, may include systems such as those disclosed in U.S. Pat. No. 7,697,972 issued April 13, 2010; U.S. Pat. No. 8,046,052 issued October 25, 201 1 ; U.S. Pat. No. 8,180,428 issued May 15, 2012; and U.S. Pat. No. 8,663,120 issued March 4, 2014, all of which are herein incorporated by reference.
  • the instrument 62 may be tracked relative to the patient 14 which may be displayed such as within the graphic representation 90 relative to selected portions including the avatar 120, the image 78, and/or an atlas 78m, as illustrated in Fig. 8.
  • the atlas 78m may be based upon the selected information, such as based on a plurality of images of patients that have been averaged and portions thereof identified.
  • the atlas 78m may be a model based on the image data from a plurality of images.
  • the atlas 78m (also referred to as a model) may have selected portions therein that are identified.
  • the atlas 78m may include the following identified locations: a right ventricle 300, a superior vena cava ostium 304, a tricuspid valve 308, a right ventricle 312, and a right ventricle septum 316.
  • the image 78 is an example of one image in a set of images that can be used to build the atlas 78m.
  • the atlas 78m may also be generated using the patient’s pre-procedure imaging like a CT or MRI.
  • the image 78 may be obtained with a selected imaging system, including the US imaging system 12, as discussed above.
  • a patient specific anatomy model may be generated using multiple 2D/3D views/planes stitched together, for example to reconstruct the endocardium of the heart.
  • a clinician may identify various relevant portions or locations within the image 78.
  • the user 18 may identify portion(s) or location(s) in the image 78 shown in the right-hand side of Fig.
  • Other locations may include endocardium or blood pool boundary may include the identification of the portions/locations within the heart. The location identification may be done manually by the user 18 and/or with assistance of an automatic location identification, such as segmenting a valve, etc.
  • Multiple images similar to the image 78 e.g., multiple similar images from a population of patients other than the patient 14, with similar identified portion(s) or location(s) may be combined to build the atlas 78m to illustrate the identified portion(s) or location(s) on a model of the imaged anatomy or heart, which model represents a typical or average patient anatomy.
  • the atlas 78m may be built only from images of the anatomy of the patient 14, and the portion(s) or location(s) of interest applied to the atlas based on identification within the images of the patient’s own anatomy, and/or based on a compilation of images from other patients’ anatomy and an estimated and/or average position of the portion(s) or location(s) of interest within the compilation of images.
  • the atlas 78m may also be morphed to actual measurements taken of the patient 14.
  • a processing module or unit such as the processor 74, may execute selected instructions to register or identify various portions in the image 78 and/or on the atlas 78m for viewing by the user 18. For example, in the atlas 78m the right ventricle septum may be identified as a selected point including a target 330.
  • the image 78 may also, therefore, have the target identified thereon 330x to illustrate a proposed or selected position of the right ventricle septum (RV septum), e.g., to serve as a target for delivery of a catheter, stylet and/or pacing lead, or a leadless pacemaker.
  • RV septum right ventricle septum
  • the predetermined portions in the model 78m may be used to assist in identifying portions of the image 78 based on one or more portions identified by the user 18 and/or appropriate image identification systems. Therefore, the model 78m may be used to assist in identifying portions within the image 78 to assist in the procedure.
  • the user may view the image 78 and identify or evaluate the target 330x prior to and/or during the procedure.
  • the image 78 may also include a plurality of images that are generated with the imaging system 12.
  • the US imaging system may also include a plurality of images that are generated with the imaging system 12.
  • the 12 may acquire images of the patient 14, including the heart 15, in a plane.
  • the plane of the image relative to the patient 14 may also change.
  • the image 78 may include a plurality of images that are illustrated in one or more planes.
  • schematic images that may represent an approximately coronal to the body co-ordinate system image 78c may be acquired and displayed and/or an image that is approximately axial to the body co-ordinate system 78ax may be displayed.
  • the two images 78ap and 78ml may both be displayed and registered to the patient 14, as discussed above.
  • the tracked position of the instrument 62 may be displayed relative to one or both of the images 78ml and 78ap.
  • both of the images 78ml and 78ap may assist in registering or relating of the images to the model 78m.
  • the user 18 may view the images, as discussed above, and identify portions or locations thereon and of the model 78m that may be used to assist in identifying various other portions.
  • the user 18 may identify the tricuspid valve 308’ and the superior vena cava entrance 304’ and the model 78m may be used to identify various other portions or locations, such as the target 330x on the RV septum 330.
  • the two images 78ap and 78ml may also assist in registration to the model 78m.
  • the identification of anatomy in the images 78 and/or one or both of the images and 78ap and 78ml may be assisted by the registration of more than one image relative to the model or atlas 78m.
  • the ultrasound imaging system 12 is registered to the patient 14, thus the images acquired with of the imaging system 12 may also be registered to the patient 14.
  • the registration of the images to the patient may be based upon the tracking of the imaging system 12, such as with the imaging system tracking device 22 and the tracking of the patient 14, such as with the patient tracking device 56.
  • the instrument 62 may be tracked relative to the patient 14 and its pose may be displayed relative to the image 78. As discussed above, the pose of the instrument 62 may also be displayed relative to the avatar 120.
  • the graphic representation 90 allows the user 18 to view the pose of the instrument 62 relative to both of the image 78 and the avatar 120.
  • the instrument 62 may be tracked in real time. Therefore, the instrument may be displayed relative to the avatar 120 and/or the image 78 to provide a real time graphical representation of the instrument 62 relative to the heart 15, based upon a graphical representation displayed relative to the image 78. In addition to viewing the tracked pose of the instrument 62, the graphical representation may show the pose of the instrument 62 over time. Therefore, the graphic representation 90 may illustrate or provide a sense of an entire path of the instrument 62 as it moves from the incision 150 in the patient 14 to the current location within the patient 14. Further, the graphic representation 90 may provide a real-time feedback to the user 18 of a movement of the instrument 62.
  • the images 78 may be acquired or based upon image data acquired at a selected time. For example, prior to the incision 150, image data may be acquired with the imaging system 12 to generate the image 78. Image 78 may thus be registered to the patient 14 and displayed with a display device 26 either alone and/or superimposed on the avatar 120. Nevertheless, the image 78 may not be a real time or current image and may be a static image. The instrument 62, however, may be tracked in real time and provide a current determination of the pose of the instrument 62. The graphic representation 90 may, therefore, be updated in real time and at a selected rate to illustrate a current position of the instrument 62 and/or a current shape or condition of the patient anatomy.
  • the display of the graphical representation may also be updated in a selected rate to provide a perception of motion of the instrument 62.
  • a refresh of the pose of the instrument 62 and of the graphic representation 90 may be 10 times a second, 30 times a second, or any appropriate rate such as about one time a second to about 120 times per second.
  • the pose of the instrument 62 may be illustrated to show a motion within the display device 26.
  • the image 78 may be displayed on the display device 26.
  • the image 78 may not change over time as the image 78 may be a static image that is acquired at a selected time, such as prior to positioning the instrument 62 within the patient 14. Nevertheless, the tracking device 60 may be tracked in real time on the device 62.
  • the graphic representation 90 may include a first or initial time graphic representation 90t1 .
  • the graphic representation 90t1 may represent a first graphic representation at a selected time, that may be arbitrarily referred to as a first or initial time.
  • a second graphic representation 90t2 may be displayed on the display device 26, as illustrated in Fig. 10B.
  • the graphic representation 90t2 may be displayed or have a selected position that is displaced a certain distance of 350 relative to the first graphic representation 90t1 .
  • the distance 350 may be any appropriate distance and may be based upon the various physical features, such as movement of a beating heart.
  • the heart 15 of the patient 14 may beat (i.e., move due to the heartbeat), as is understood by one skilled in the art. Therefore, the graphic representation 90t2, including a terminal end 90x, may be shown in different positions based upon movements of the instrument 62. For example, as one skilled in the art will understand, when the terminal end of the instrument 62 contacts the RV septum 316, the terminal end of the instrument 62 may move due to the beating of the heart 15.
  • the instrument 62 may appear to move in real time.
  • Real-time tracking of the instrument 62 with the tracking device and 60 may allow for a representation of the movement of the instrument 62 to be displayed on the display device 26.
  • the movement of the instrument may be updated and displayed at any appropriate rate to provide a perception of movement to the user 18.
  • the user 18 may be provided with additional information regarding a position of the instrument 62, such as being in contact with of the RV septum 316.
  • the process 370 may be a process that is executed by one or more of the processing units, such as processing unit 74.
  • the method 370 may begin at start Block 374.
  • the process 370 may determine a first position or pose of the instrument in Block 378. Determining the first pose of the instrument may be performed with the navigation system, as discussed above.
  • the pose of the instrument 62 may be known relative to the patient 14, or portions thereof such as the heart 15, in light of the tracking of the patient with the patient tracker 58.
  • an output for display of the determined pose of the instrument may be made in Block 382.
  • the output of the determined pose may be based upon the registration of the image 78 and/or the avatar 120 to the physical space defined by the patient 14, as discussed above.
  • the output may display the graphic representation 90 at a first position, for example of the first pose 90t1 , as illustrated in Fig. 10A.
  • a wait or pause for a selected time may occur in Block 384.
  • the wait for a selected time may be any appropriate time, such as about one second to about 120 seconds.
  • the wait time may be any appropriate time to assist in illustrating a motion to the user 18, as discussed above.
  • the second pose may be determined in Block 388.
  • the second pose may be determined in the same manner as the first pose was determined in Block 384.
  • An output for display of the second pose may be made in Block 392 to display the graphical representation of the instrument 62 at a second time, such as the graphic representation 90t2. As noted above, therefore, the pose of the instrument 62 may be displayed over a series of the events and/or time to show movement to the user 18.
  • the process 370 may provide an illustration to the user 18 for an appropriate period of time. If a tracking is not to be continued, a NO path 400 may be followed to end the process in Block 404.
  • the process 370 may be used to illustrate and determine a pose of the instrument 62 over time for illustration relative to the image 78, as illustrated in Fig. 10A and 10B.
  • the display 26 may display the image of the patient 14 that is acquired with the imaging system 12. As discussed above, the image 78 may be an image that is acquired in a plane from the ultrasound housing 16.
  • rotation of the image may change a perspective of the image 78.
  • the image may be viewed over the plane 78f or from an edge 78e.
  • the edge image 78e may allow for a viewing and understanding of a perspective of the instrument 62, as illustrated by the graphic representation 90, relative to a plane of the image.
  • the image acquired with the imaging system 12 may be a static image that is acquired with the imaging system 12. It is understood, however, that a live or current image may also be acquired and displayed. Nevertheless, the position of the instrument 62 relative to the plane of the image may be illustrated by rotating the image, such as showing it from an edge as illustrated in the edge image 78e.
  • the distance of the instrument 62 that is within the plane of the image may also be illustrated.
  • a first and second distance icon 410, 412 may be included or illustrated superimposed on both of the images 78e, 78f.
  • the distance icons may be used to illustrate the distance that the instrument 62 is in the plane of the image.
  • the user 18 may also understand that the portion of the instrument that is within the plane of the image and that other portions are not within the plane of the image.
  • the distance icon 410, 412 may be sized depending upon the length or portion of the instrument 62 that is within the plane of the image 78. Therefore, the first distance icon 410 may have a first distance or size 420.
  • the second distance icon 412 may have a second distance or size 424.
  • the two dimensions 420, 424 may be different or the same depending upon the amount of the instrument or length of the instrument that is within the plane of the image 78. If the instrument has a greater length within the plane, the distance icons 410, 412 may have greater dimensions. If any dimension of the instrument is smaller in the plane, then the dimension icons may be smaller.
  • the dimension icons 410, 412 may be displayed, such as superimposed on, the image 78 to represent the distance of the instrument within the image plane. This may allow the user 18 to understand or determine the distance of the instrument 62 within a plane of the imaging system and/or assist in moving of the imaging system to acquire a different perspective than in the image 78 relative to the instrument 62.
  • the display 26 may display the image of the patient 14 that is acquired with the imaging system 12 at more than one location. In other words, more than one image may be displayed for each position of the imaging system 12 at which image data is acquired. Moreover, one or more image data portions may be merged and/or used in a reconstruction of a displayed image. As discussed above and herein, all of the images and/or the reconstructed image may be registered to the patient 14 due at least to the tracking of the imaging system 12.
  • the image 78 may be an image that is acquired in a plane from the ultrasound housing 16. Therefore, rotation of the image may change a perspective of the image 78.
  • the image may be viewed as acquired at a first location (e.g., inferior from the suprasternal notch) as over a plane 78f ’ or from an edge 78e’. These two views may be able to provide information to the user, as discussed above including location and/or dimension that the instrument graphic representations 90 passes through the plane 410’, 412’, 420’, 424’.
  • the image may be viewed as acquired at a second location (e.g., between selected ribs) as over a plane 78f” or from an edge 78e”.
  • a second location e.g., between selected ribs
  • These two views may be able to also provide information to the user, as discussed above including location and/or dimension that the instrument graphic representations 90 passes through the plane 410”, 412”, 420”, 424”.
  • the graphic representations 90 is illustrated to pass through the planes at the different positions and/or dimensions 41 O’, 412’, 420’, 424’, 410”, 412”, 420”, 424” and this information is displayed for the user.
  • any appropriate number of location images may be acquired and/or displayed and two is merely exemplary.
  • the edge image may allow for a viewing and understanding of a perspective of the instrument 62, as illustrated by the graphic representation 90, relative to a plane of the image.
  • the image acquired with the imaging system 12 may be a static image that is acquired with the imaging system 12. It is understood, however, that a live or current image may also be acquired and displayed. Nevertheless, the position of the instrument 62 relative to the plane of the image may be illustrated by rotating the image, such as showing it from an edge as illustrated in the edge image.
  • the instrument 62 may be tracked with any appropriate tracking system.
  • tracking and/or navigation systems include those disclosed above and those disclosed in U.S. Pat. No. 7,697,972 issued April 13,
  • the navigation system may allow for a determination of a pose of the instrument 62 in a physical location and/or orientation of the instrument 62 in the physical space.
  • the tracked pose may include a location which includes X, Y, and Z axis positions and/or orientation which may include yaw, pitch, and roll orientations in space.
  • a tracking and navigation system may also include one or more permanent magnets and/or magnetometers.
  • the navigation processing system 70 is further configured to generate the display 26 of Fig. 13.
  • the display 26 shows a probe housing icon 16' of the housing 16 of the imaging system 12. That is, the icon 16’ is displayed relative to, such as superimposed on, the avatar 120. This allows the user to understand the pose of the probe 16 relative to the subject and the image data acquired with the probe 16.
  • FIG. 1 Shown emanating from the probe housing icon 16' is a field of view 24 of the imaging system 12.
  • the portion of the anatomy within the field of view 24 captured by the imaging system 12 is shown as the ultrasound image 78.
  • a heart image or model 110 is within the field of view 24.
  • the heart image would be an actual ultrasound image captured by the imaging system 12.
  • a 3D model would be a model created from segmented ultrasound images or from a registered CT or MRI scan.
  • the location of the field of view 24 on the display 26 is determined based on the pose of the housing 16, as determined by tracking the position of the tracking device 22. Showing the field of view 24 on the display 26 assists the operator of the imaging system 12 with positioning the housing 16 to capture a desired portion of the anatomy.
  • the user can see on the avatar 120 the portion of the anatomy within the field of view 24, which is displayed as the US image 78. If an area of interest is not within the field of view, the user can use the avatar 120 as a guide for maneuvering the housing 16 until the area of interest is within the field of view 24 and displayed in the ultrasound image 78.
  • the US image 78 of Fig. 13 is captured when the probe 16 is in the orientation of the probe housing icon 16' shown in Fig. 13.
  • the user can use the avatar 120, the heart image or model 1 10, the housing icon 16', and the field of view 24 as guides for maneuvering the actual housing 16 to bring the area of interest into the field of view 24.
  • the user can maneuver the actual housing 16 until the field of view 24 is at the mid-papillary level of the heart image or model 1 10. This is the case in Fig. 14, where the housing 16 has been maneuvered such that the field of view 24B extends through the mid-papillary level of the heart image or model 1 10, thereby resulting in the capture of US image 78B at the mid-papillary level of the heart.
  • the user or selected processor executing instructions may view and understand the a current position of the probe and move the probe to a new position based on the current position.
  • the navigation processing system 70 can also be configured to generate the display 26 of Fig. 14.
  • the display 26 includes three panels.
  • the center panel shows the avatar 120 displaying the subject’s heart, or other relevant anatomy, as the heart image or model 1 10.
  • the probe housing icon 16' Also shown on the center panel is the probe housing icon 16', a first field of view 24A, and a second field of view 24B.
  • the first field of view 24A corresponds to the US image 78A on the left of the display 26.
  • the second field of view 24B corresponds to the US image 78B on the right side of the display 26.
  • the first field of view 24A and the second field of view 24B may be captured by two different housings 16, or by the same housing 16 at different times.
  • the atlas model 78M may be generated by capturing a plurality of US images of the subject using the imaging system 12.
  • the images collected of the subject such as the heart, may also be used to morph an atlas of the heart of the current subject.
  • a user must thus maneuver the housing 16 to capture an entirety of the anatomy, such as an entirety of the heart. It is understood, however, that collecting a sufficient amount of image data of the subject may be appropriate to morph an atlas or generate an atlas.
  • the navigation processing system 70 may be configured to generate the display 26 of FIG. 15, which includes four panels displaying different sides of the heart image or model 1 10.
  • the heart image or model 1 10 is displayed in any suitable manner that distinguishes areas of the heart that have been scanned by the imaging system 12 versus areas of the heart that have not been scanned by the imaging system 12. During the scanning process, the navigation system 70 tracks movement of the housing 16 relative to the anatomy.
  • first portions 1 12 of the heart image or model 1 10, which are illustrated with solid lines, have already been imaged by the imaging system 12.
  • the navigation processing system 70 tracks the position of the housing 16 relative to the anatomy, in this case the heart, to identify Second portions 114 of the heart image or model 1 10, which are illustrated in phantom, have yet to be imaged.
  • the navigation processing system 70 is configured to update the display 26 as the heart is scanned to reflect the additional scanning. For example, and with respect to the images of Fig. 15, as the atriums of the heart are scanned the navigation processing system 70 is configured to update the heart image or model 110 to convert the atriums from second portions 1 14 in phantom to first portions 1 12.
  • the navigation system 70 tracks movement of the housing 16 relative to the heart, or other portions of the anatomy being scanned, to identify portions of the heart that have been scanned versus portions that have not been scanned.
  • the navigation system 70 continuously updates the images of Fig. 15 as described above to distinguish the portions of the heart that have been scanned versus the portions that have yet to be scanned based on tracked movement of the housing 16 relative to the heart.
  • the imaging system 12, including the ultrasound probe, may be moved relative to the subject to acquire an appropriate amount of image data, as discussed above.
  • the imaging and/or navigation processing system may be used to identify portions of the heart that have not been imaged or not imaged enough to generate a selected feature or type of image for display.
  • the image data acquired may be used to generate an image for analysis of the subject.
  • a three-dimensional or volumetric image of a portion of the subject, such as the heart, may be generated based upon the image data acquired with of the imaging system 12.
  • the ultrasound probe generates data in a plane. Planes may intersect each other when generated or collected with the ultrasound probe.
  • the ultrasound probe is tracked during acquisition of the image data planes that may move and change relative to one another, such as being orthogonal.
  • the image data acquired at different perspectives (e.g., different planes) that are known relative to one another allows a reconstruction to be made.
  • the reconstruction may be of the heart as the image 110.
  • the reconstruction may be a three-dimensional reconstruction based upon the image data collected with the ultrasound probe and the processor may illustrate portions of the heart needed to be imaged to reconstruct the 3D image, as noted above.
  • Reconstruction techniques may include, but are not limited to, the following: computed tomography, specifically ultrasound computed tomography; and NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis, Ben Mildenhall, Pratul P. Srinivasan, Matthew Tancik, Jonathan T. Barron, Ravi Ramamoorthi, Ren Ng (August 3, 2020)
  • the navigation processing system 70 is configured to generate a display 26 including a sensor icon 160 arranged relative to the avatar 120.
  • the position of the sensor icon 160 represents the position of any of the localizers relative to the subject 14, such as any of the EM localizers 50, 55, or 58.
  • a user may understand by viewing the avatar 120 the pose of the EM localizers 50, 55, or 58.
  • the EM localizers 50, 55, or 58 may emit a field that induces a current in a tracking device, such as the tracking device 22.
  • the display 26 also includes a range icon 162, which identifies the outer boundaries of the range of the particular localizer displayed.
  • the range icon 162 identifies that extent (e.g., area or volume), in which the field emitted by the EM localizer is effective for navigation. Similarly, a field of view of an optical localizer may be displayed as the range icon 162. The range icon may be displayed relative to the subject aviator 120.
  • Displaying the sensor icon 160 and the range icon 162 facilitates positioning of the subject 14 and/or the localizer 50, 55, or 58 to capture an area of the body of interest. That is, the range icon 162 illustrates that extent to which the EM localizer 50, 55, or 58 is effective. As the avatar 120 illustrates the subject to be imaged and/or have tracking devices tracked relative thereto, the range illustrates the area in which a tracking device may be effectively tracked. Thus, the user may view the range icon 12 and decide if the volume included is effective for the procedure or not.
  • the localizer is registered in any suitable manner. For example, registration may be carried out with a reference tracker on the subject’s body at a predefined location, such as the reference tracker 56. That is, the localizers 50, 55, or 58 may sense a reference marker placed on the subject at a known anatomical location and the localizer may be used to determine the pose of the localizer relative thereto. Further, the localizer range 162 may be relative to a second localizer that is tracked with a first localizer.
  • Fig. 17 illustrates another exemplary display 26 in accordance with the present disclosure. The display 26 of FIG. 17 displays the avatar 120 including a first respiratory sensor icon 170 and a second respiratory icon 172 relative to the avatar 120.
  • the first respiratory icon 170 represents the position on the subject’s chest of a first respiratory sensor including an EM tracker.
  • the second respiratory icon 172 represents the position on the subject’s back of a second respiratory sensor including an EM tracker.
  • the trackers may be placed in the same general vicinity by convention. One region of high motion (e.g., substernal/subxiphoid region) and one region of lesser motion (upper back). Or alternatively the patch may be placed and the user selects from a generic menu of where it was placed from an array of options that would make sense (regions of the upper torso). It is typically desirable to arrange the first and the second respiratory sensors so that they are aligned with respect to the orientation of the avatar 120 in FIG. 17.
  • the relative positions of the first and the second respiratory sensor icons 170, 172 in Fig. 17 allows medical personnel to determine whether the actual respiratory sensors are aligned.
  • the display 26 also includes a respiration meter 180, which includes additional first respiratory icons 170' and second respiratory icons 172'. During respiration, the icons 170', 172' move relative to each other to visually illustrate respiration. This allows monitoring for the phase and depth of respiration, identifying abnormalities with monitoring (e.g., tracker becomes disconnected or isn’t well attached), and/or abnormalities in the regularity of the breathing of the patient, which may result in inconsistent tracking or imaging. [0117] Example embodiments are provided so that this disclosure will be thorough and will fully convey the scope to those who are skilled in the art.
  • the described techniques may be implemented in hardware, software, firmware, or any combination thereof. If implemented in software, the functions may be stored as one or more instructions or code on a computer-readable medium and executed by a hardware-based processing unit that may also be referred to as a processor.
  • Computer-readable media may include non-transitory computer-readable media, which corresponds to a tangible medium such as data storage media (e.g., RAM, ROM, EEPROM, flash memory, or any other medium that can be used to store desired program code in the form of instructions or data structures and that can be accessed by a computer).
  • processors or processor modules such as one or more digital signal processors (DSPs), general purpose microprocessors, application specific integrated circuits (ASICs), field programmable logic arrays (FPGAs), or other equivalent integrated or discrete logic circuitry.
  • DSPs digital signal processors
  • ASICs application specific integrated circuits
  • FPGAs field programmable logic arrays
  • processors or processor modules may refer to any of the foregoing structure or any other physical structure suitable for implementation of the described techniques. Also, the techniques could be fully implemented in one or more circuits or logic elements.
  • Example 1 A system configured to assist a user with navigating an instrument relative to a subject, the system comprising: an imaging system configured to image an anatomy, the imaging system including a probe maneuverable by the user; a tracking system including a localizer, a probe tracker configured to be mounted to the probe, and a subject tracker configured to be mounted to the subject; a display device including a display screen; and a processor configured to: generate an avatar of the subject and display the avatar in a first area of the display screen; display an anatomy icon of the anatomy on the avatar in the first area of the display screen; display a probe icon and a field of view icon of the probe in the first area of the display screen at a position relative to the avatar corresponding to position of the probe relative to the anatomy; and display a first image of the anatomy captured by the probe in a second area of the display screen, the anatomy in the first image corresponding to the anatomy icon within the field
  • Example 2 The system of example 1 , wherein the imaging system is an ultrasound imaging system, and the probe is an ultrasound probe.
  • Example s The system of example 1 , wherein the anatomy includes a heart.
  • Example 4 The system of example 1 , wherein the processor is configured to display the first image at a first orientation or a second orientation that is different than the first orientation without the probe being moved.
  • Example s The system of example 1 , wherein the processor is further configured to: display a second image of the anatomy in a third area of the display screen, the second image shows a different area of the anatomy than the first image.
  • Example 6 The system of example 5, wherein the second image is captured by the probe before or after the probe captures the first image; and wherein the processor is configured to display another field of view icon on the display corresponding to the second image.
  • Example 7 A system configured to assist a user with navigating an instrument relative to a subject, the system comprising: an imaging system configured to image an anatomy, the imaging system including a probe maneuverable by the user; a tracking system including a localizer, a probe tracker configured to be mounted to the probe, and a subject tracker configured to be mounted to the subject; a display device including a display screen; and a processor configured to: display a first anatomy icon of the anatomy in a first area of the display screen; and distinguish visually on the first anatomy icon a first portion of the anatomy that has been imaged by the probe and a second portion of the anatomy that has not been imaged by the probe.
  • Example s The system of example 7, wherein the processor is further configured to display a second anatomy icon of the anatomy in a second area of the display screen, the second anatomy icon shows a different angle of the anatomy as compared to the first anatomy icon.
  • Example 9 The system of example 8, wherein the processor is further configured to display a third anatomy icon of the anatomy in a third area of the display screen, and display a fourth anatomy icon of the anatomy in a fourth area of the display screen; wherein each of the first anatomy icon, the second anatomy icon, the third anatomy icon, and the fourth anatomy icon are orientated at different angles; and wherein the processor is configured to distinguish visually on each of the second anatomy icon, the third anatomy icon, and the fourth anatomy icon the first portion of the anatomy that has been imaged by the probe and the second portion of the anatomy that has not been imaged by the probe.
  • Example 10 The system of example 7, wherein the anatomy is a heart.
  • Example 1 A system configured to assist a user with navigating an instrument relative to a subject, the system comprising: a tracking system including a localizer and a subject tracker configured to be mounted to the subject; a display device including a display screen; and a processor configured to: generate an avatar of the subject and display the avatar on the display screen; display an anatomy icon of an anatomy on the avatar; and display a localizer icon and a range icon of the localizer on the display screen at a position relative to the avatar corresponding to position of the localizer relative to the subject.
  • a tracking system including a localizer and a subject tracker configured to be mounted to the subject
  • a display device including a display screen
  • a processor configured to: generate an avatar of the subject and display the avatar on the display screen; display an anatomy icon of an anatomy on the avatar; and display a localizer icon and a range icon of the localizer on the display screen at a position relative to the avatar corresponding to position of the localizer relative to the subject.
  • Example 12 The system of example 1 1 , wherein the localizer is an electromagnetic localizer.
  • Example 13 The system of example 12, wherein the electromagnetic localizer is suspended over an anterior side of the subject.’
  • Example 14 The system of example 12, wherein the electromagnetic localizer is at a posterior side of the subject.
  • Example 15 The system of example 12, wherein the electromagnetic localizer is mounted at a side of the subject.
  • Example 16 A system configured to assist a user with navigating an instrument relative to a subject, the system comprising: a respiratory monitoring system including a first respiration sensor and a second respiration sensor; a display device including a display screen; and a processor configured to: generate an avatar of the subject and display the avatar on the display screen; display a first respiration sensor icon on the avatar at a first location corresponding to placement of the first respiration sensor on the subject; and display a second respiration sensor icon on the avatar corresponding to placement of the second respiration sensor on the subject.
  • Example 17 The system of example 16, wherein the first respiration sensor and the second respiration sensor include electromagnetic localizers.
  • Example 18 The system of example 16, wherein the processor is further configured to display on the display screen a respiration meter including a first meter icon and a second meter icon.
  • Example 19 The system of example 18, wherein the processor is configured to move the first meter and the second meter icon on the display screen in a manner corresponding to movement of the first respiration sensor and the second respiration sensor during respiration.

Landscapes

  • Health & Medical Sciences (AREA)
  • Engineering & Computer Science (AREA)
  • Surgery (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Robotics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Biomedical Technology (AREA)
  • Medical Informatics (AREA)
  • Molecular Biology (AREA)
  • Animal Behavior & Ethology (AREA)
  • General Health & Medical Sciences (AREA)
  • Public Health (AREA)
  • Veterinary Medicine (AREA)
  • Human Computer Interaction (AREA)
  • Measuring And Recording Apparatus For Diagnosis (AREA)

Abstract

A system and method to provide output regarding an imaging system. The output includes, in various embodiments, a display illustrating portions features and portions selected by a user. The imaging system may include a user moveable and/or holdable imaging system.

Description

SYSTEM AND METHOD FOR ILLUSTRATING A POSE OF AN OBJECT
[0001] This application claims the benefit of U.S. Provisional Patent Application, Serial No. 63/495,951 , filed April 13, 2023, and U.S. Utility Patent Application, Serial No. 18/630,474, filed April 9, 2024, the entire contents of each are incorporated herein by reference.
CROSS-REFERENCE TO RELATED APPLICATIONS
[0002] The subject application includes subject matter related to the subject matter disclosed in U.S. Provisional Patent Applications 63/420,274 filed on October 28, 2022; 63/420,315 filed on October 28, 2022; and 63/420,328 filed on October 28, 2022, all of which are incorporated herein by reference.
FIELD
[0003] The present disclosure relates to acquisition of image data of a subject, and particularly to acquisition and display of image data collected from a tracked imaging system.
BACKGROUND
[0004] This section provides background information related to the present disclosure which is not necessarily prior art.
[0005] An imaging system can be used to image various portions of a subject. The subject can include a patient, such as a human patient. The portions selected to be imaged can be internal portions that are covered by skin or other tissue. A location of imaged portions of the subject may be selected to be known within the imaging system. The locations can be defined or established relative to instruments placed in the subject (e.g., a location of a heart wall relative to a catheter) or a location of the imaged portion relative to the instrument acquiring the image data.
[0006] An imaging modality may be used to generate image data that are used to render images of an interior of the subject. For example, x-rays may be used to generate image data of an interior of a subject. The images may be displayed for viewing by a user.
SUMMARY
[0007] This section provides a general summary of the disclosure, and is not a comprehensive disclosure of its full scope or all of its features.
[0008] The subject disclosure relates to illustrating a position, including a pose of an instrument relative to a patient. The patient may be a human or nonhuman patient. Further, the patient or subject may be a non-living object where the location of an instrument within the subject is selected to be tracked.
[0009] A representation of the patient may be fit or morphed to illustrate or represent a size of the patient. Therefore, relative portions of the patient, such as relative portions of the anatomy, are determined or known relative to one another based upon the patient that is the current patient during a procedure. An avatar or indicia of the patient may be generated or altered to a specific patient substantially in real time during a procedure.
[0010] An instrument can be tracked relative to the patient. A representation of the instrument may be superimposed on the avatar of the patient. The avatar of the patient may provide a visual illustration of various portions of the anatomy of the patient without requiring image data being acquired of the patient. The display being viewed by the user may be generated without requiring image data of the patient, such as fluoroscopy image data.
[0011] Further areas of applicability will become apparent from the description provided herein. The description and specific examples in this summary are intended for purposes of illustration only and are not intended to limit the scope of the present disclosure.
DRAWINGS
[0012] The drawings described herein are for illustrative purposes only of selected embodiments, not all possible implementations, and are not intended to limit the scope of the present disclosure.
[0013] Fig. 1 is an environmental view of an exemplary imaging and navigation system;
[0014] Fig. 2A is a screen shot of a representation of an instrument at a tracked pose relative to an avatar of a subject, the screen shot including an image taken at a first orientation;
[0015] Fig. 2B is similar to Fig. 2A, but with the image at a second orientation that is about 90° relative to the first orientation;
[0016] Fig. 3 is a schematic view of a patient in an operating theater;
[0017] Fig. 4 is a schematic view of a user and a patient in an operating theater, according to various embodiments;
[0018] Fig. 5 is a detailed view of an instrument being moved relative to a patient; [0019] Fig. 6 is a screen shot of a display illustrating a position of an instrument relative to an avatar;
[0020] Fig. 7 is a flowchart of a method, according to various embodiments;
[0021] Fig. 8 is a schematic illustration of an atlas and an image;
[0022] Fig. 9 is an illustration of images in two planes of a heart, according to various embodiments including ultrasound;
[0023] Figs. 10A and 10B illustrate movement of a tracked instrument over time;
[0024] Fig. 11 is a flowchart illustrating a method of showing movement of a tracked instrument over time, according to various embodiments;
[0025] Fig. 12A illustrates a tracked pose of an instrument relative to an image in a plane in two orientations, according to various embodiments;
[0026] Fig. 12B illustrates a tracked pose of an instrument relative to an image in a plane in two orientations, according to various embodiments;
[0027] Fig. 13 illustrates an exemplary screen shot with right and left panels, on the right is an avatar of a subject with a housing of an imaging device and a field of view of the imaging device overlaid on the avatar, on the left is an image captured by the imaging device;
[0028] Fig. 14 illustrates an exemplary screen shot with right, left, and center panels, the center panel includes an avatar of a subject with a housing of an imaging device overlaid on the avatar, a first field of view in the center panel corresponds to an image captured by the imaging device shown in the left panel and a second field of view in the center panel corresponds to an image captured by the imaging device shown in the right panel;
[0029] Fig. 15 illustrates an exemplary screen shot with four panels, each panel shows a different orientation of a heart icon representing a heart being scanned with an imaging device, areas of the heart that have been scanned are displayed with different shading, coloring, etc. to distinguish from areas of the heart that have not yet been scanned;
[0030] Fig. 16 illustrates an exemplary screen shot with a sensor icon and a range icon overlaid on the avatar to identify a position and range of a sensor; and
[0031] Fig. 17 illustrates an exemplary screen shot with a left panel including a first respiration sensor icon and a second respiration sensor icon arranged on the avatar, and a right panel including a graphical representation of movement between first and second respiration sensors on a subject’s body.
[0032] Corresponding reference numerals indicate corresponding parts throughout the several views of the drawings.
DETAILED DESCRIPTION
[0033] Example embodiments will now be described more fully with reference to the accompanying drawings.
[0034] Fig. 1 is a diagram illustrating an overview of a navigation system 10 that can be used for various procedures. The navigation system 10 can be used to track the location of an item, such as an implant or an instrument, and at least one imaging system 12 relative to a subject, such as a subject 14 that may include a human patient and/or other living or non-living subject. It should be noted that the navigation system 10 may be used to navigate any type of instrument, implant, or delivery system, including: catheters, stylets, leads for cardiac rhythm management devices such as pacemakers, leadless pacemakers and delivery systems therefor, guide wires, arthroscopic systems, ablation instruments, stents, orthopedic implants, spinal implants, deep brain stimulation (DBS) probes, mechanical parts, Transesophageal Echocardiography (TEE), intra-cardiac echocardiography (ICE), etc. Non-human or non-surgical procedures may also use the navigation system 10 to track a non-surgical or non-human intervention of the instrument or imaging device. Moreover, the instruments may be used to navigate or map any region of the body. The navigation system 10 and the various tracked items may be used in any appropriate procedure, such as one that is generally minimally invasive or an open procedure.
[0035] The navigation system 10 can interface with or integrally include an imaging system 12 that is used to acquire pre-operative, intra-operative, or postoperative, or real-time image data of the patient 14. For example, the imaging system 12 can be an ultrasound imaging system (as discussed further herein) that has a tracking device 22 attached thereto. The imaging system 12 may also include any suitable CT system, x-ray system, C-arm, O-arm® imaging system, magnetic resonance imaging system, etc. The tracking device 22 may be tracked with the tracking system to determine a pose of the imaging system 12. The pose may include an orientation (e.g., three or more degrees of orientation (e.g., yaw, pitch, and roll) and/or a position (e.g., three degrees of freedom in physical space (e.g., x-axis, y-axis, and z-axis). The tracking system 10 may further determine appropriate pose information regarding the tracking device 22. The pose of the imaging system 12 can then be determined based on the tracked pose of the tracking device 22. The imaging system 12 may be used to generate image data to provide images for viewing with a selected display device 26, which may any appropriate display device including an augmented and/or virtual reality display device worn/used by the user 18 such as those disclosed in U.S. Pat. App. Pub. No. US2018/0078316A1 , incorporated herein by reference.
[0036] It will be understood any appropriate subject can be imaged and any appropriate procedure may be performed relative to the subject. The navigation system 10 can be used to track various tracking devices, as discussed herein, to determine locations of the patient 14. The tracked poses of the patient 14 can be used to determine or select images for display to be used with the navigation system 10. The initial discussion, however, is directed to the navigation system 10 and the exemplary imaging system 12.
[0037] In the example shown, the imaging system 12 includes an ultrasound (US) imaging system with an US housing 16 that is held by a user 18 while collecting image data of the subject 14. It will be understood, however, that the US housing 16 can also be held by a stand or robotic system while collecting image data. The US housing 16 and included transducer can be any appropriate US imaging system 12, such as the M-TURBO® sold by SonoSite, Inc. having a place of business at Bothell, Washington. Associated with, such as attached directly to or molded into, the US housing 16 or the US transducer housed within the housing 16 is at least one imaging system tracking device 22. The tracking device 22 may be any appropriate tracking device such as an electromagnetic (EM) tracking device and/or an optical tracking device. The tracking devices can be used together (e.g., to provide redundant tracking information) or separately. Also, only one of the two tracking devices may be present. It will also be understood that various other tracking devices can be associated with the US housing 16, as discussed herein, including acoustic, ultrasound, radar, electrical impedance, and other tracking devices. Also, the tracking device 22 can include linkages or a robotic portion that can determine a location relative to a reference frame. Any suitable tracking device of any suitable tracking system may be used, such as, tracking devices and systems as understood by one skilled in the art.
[0038] In various embodiments, a secondary imaging system may also be presented and/or used to generated image data of the patient 14. The second imaging system may include an O-arm® imaging device sold by Medtronic Navigation, Inc. having a place of business in Louisville, Colorado, USA, a C-arm imaging system, or other appropriate imaging system. The second imaging system can include those disclosed in U.S. Pat. Nos. 7,188,998; 7,108,421 ; 7,106,825; 7,001 ,045; and 6,940,941 ; all of which are incorporated herein by reference. The second imaging system, however, need not be present and the imaging system 12, which may be only the US transducer, may be the only imaging system to generate images to be displayed. As discussed herein, various other image data may also be displayed such as an avatar 120 of the patient 14. [0039] The imaging system 12 may be tracked, as discussed above. Various tracking systems may include and/or require calibration of the imaging system. Thus, the pose of the tracking device 22 relative to a plane of the US imaging system may be determined and/or known. Various systems and methods are disclosed in U.S. Patent Nos. 6,379,302; 6,669,635; 6,968,224; 7,085,400; 7,831 ,082; 8,320,653; 8,811 ,662; and 9,138,204 all of which are incorporated herein by reference.
[0040] As illustrated in Figs. 1 , 3, and 4, the patient 14 can be fixed in a selected pose relative to a selected object, such as onto an operating table 40, but is not required to be fixed to the table 40. The table 40 can include a plurality of straps 42 or other fixation devices. The straps 42 can be secured around the patient 14 to fix the patient 14 relative to the table 40. Various apparatuses may be used to position the patient 40 in a static position on the operating table 40. Examples of such patient positioning devices are set forth in commonly assigned U.S. Pat. App. No. 10/405,068, published as U.S. Pat. App. Pub. No. 2004/0199072 on October 7, 2004, entitled “An Integrated Electromagnetic Navigation And Patient Positioning Device”, filed April 1 , 2003 which is hereby incorporated by reference. Other known apparatuses may include a Mayfield® clamp.
[0041] The navigation system 10 includes at least one tracking system. The tracking system can include at least one localizer. In one example, the tracking system can include an electromagnetic (EM) localizer 50. The tracking system can be used to track instruments relative to the patient 14 or within a navigation space. The navigation system 10 can use image data from the imaging system 12 and information from the tracking system to illustrate locations of the tracked instruments, as discussed herein. The tracking system can also include a plurality of types of tracking systems including an optical localizer 52 in addition to and/or in place of the EM localizer 50. When the EM localizer 50 is used, the EM localizer can communicate with or through a localizer communication 54 that may be wired or wireless. The EM localizer 50 may also include and/or take the form of an alternative pad or flat EM localizer 55.
[0042] The optical localizer 52 and the EM localizer 50 can be used together to track multiple instruments or used together to redundantly track the same instrument. Various tracking devices, including those discussed further herein, can be tracked and the information can be used by the navigation system 10 to allow for an output system to output, such as a display device to display, a position of an item. Briefly, tracking devices can include a patient or reference tracking device 56, also known as a dynamic reference frame, to track the patient 14, an instrument tracking device 60 to track an instrument 62, and/or other appropriate tracking devices for one or more portions. Patient or reference tracking device 56 and instrument tracking device 60 may include those disclosed in U.S. Pat. Nos. 8,060,185 and 8,644,907, both incorporated herein by refence. The tracking devices allow selected portions of the operating theater to be tracked relative to one another with the appropriate tracking system, including the optical localizer 52 and/or the EM localizer 50. The reference tracking device 56 can also or alternatively be positioned on an instrument and positioned within the patient 14, such as within a heart 15 of the patient 14.
[0043] It will be understood that any of the tracking devices 22, 56, 60 can be optical or EM tracking devices, or both, depending upon the tracking localizer used to track the respective tracking devices. It will be further understood that any appropriate tracking system can be used with the navigation system 10. Alternative tracking systems can include radar tracking systems, acoustic tracking systems, ultrasound tracking systems, electrical impedance tracking systems, radio frequency beacon, and the like. Exemplary tracking systems include those disclosed in U.S. Pat. Nos. 7,676,268; 8,532,734; and 8,494,608, all incorporated herein by reference. Each of the different tracking systems can be respective different tracking devices and localizers operable with the respective tracking modalities. Also, the different tracking modalities can be used simultaneously as long as they do not interfere with each other (e.g., an opaque member blocks a camera view of the optical localizer 52).
[0044] An exemplary EM tracking system can include the STEALTHSTATION® AXIEM™ Navigation System, sold by Medtronic Navigation, Inc. having a place of business in Louisville, Colorado. Exemplary tracking systems are also disclosed in U.S. Patent No. 7,751 ,865, issued July 6, 2010 and entitled "METHOD AND APPARATUS FOR SURGICAL NAVIGATION"; U.S. Patent No. 5,913,820, titled “Position Location System,” issued June 22, 1999 and U.S. Patent No. 5,592,939, titled “Method and System for Navigating a Catheter Probe,” issued January 14, 1997, all herein incorporated by reference.
[0045] Further, for EM tracking systems it may be necessary to provide shielding or distortion compensation systems to shield or compensate for distortions in the EM field generated by the EM localizer 50. Exemplary shielding systems include those in U.S. Pat. No. 7,797,032, issued on September 14, 2010 and U.S. Pat. No. 6,747,539, issued on June 8, 2004; distortion compensation systems can include those disclosed in U.S. Pat. No. 10/649,214, filed on January 9, 2004, published as U.S. Pat. App. Pub. No. 2004/01 16803, all of which are incorporated herein by reference. [0046] With an EM tracking system, the EM localizer 50 and the various tracking devices can communicate through an EM controller. The EM controller can include various amplifiers, filters, electrical isolation, and other systems. The EM controller can also control the coils of the EM localizer 50 to either emit or receive an EM field for tracking. A wireless communications channel, however, such as that disclosed in U.S. Patent No. 6,474,341 , entitled “Surgical Communication Power System,” issued November 5, 2002, herein incorporated by reference, can be used as opposed to being coupled directly to the EM controller. The EM controller can be incorporated into a navigation processing system 70.
[0047] It will be understood that the tracking system may also be or include any appropriate tracking system, including a STEALTHSTATION® TRIA®, TREON®, and/or S7™ Navigation System having an optical localizer, similar to the optical localizer 52, sold by Medtronic Navigation, Inc. having a place of business in Louisville, Colorado. Further, alternative tracking systems are disclosed in U.S. Patent No. 5,983,126, to Wittkampf et al. titled “Catheter Location System and Method,” issued November 9, 1999, which is hereby incorporated by reference. Other tracking systems include an acoustic, radiation, radar, electrical impedance, etc. tracking or navigation systems.
[0048] The navigation system 10 can include a navigation processing unit or module 74 that can communicate or include a navigation memory 76, which may be included in the navigation processing system 70. The navigation processing system 70 may further include a display device 77. The navigation processing unit 74 can include a processor and memory (e.g., microprocessor, a central processing unit, etc.). In various embodiments, the navigation processing unit may execute instructions to determine one or more poses of the tracking devices based on signals from the tracking devices. The navigation processing unit 74 can receive information, including image data, from the imaging system 12 and tracking information from the tracking systems, including the respective tracking devices and/or the localizers 50, 54. Image data can be displayed as an image 78 on the display device 26. The display device may be separate from and/or integrated into the navigation system, thus, the display device 26 may include the display device 77. The navigation system 70 can include appropriate input devices, such as a keyboard 84. It will be understood that other appropriate input devices can be included, such as a mouse, a foot pedal 88 or the like which can be used separately or in combination. Also, all of the disclosed processing units or systems can be a single processor module (e.g., a single central processing chip) that can execute different instructions to perform different tasks. [0049] An image processing unit or module can process image data from the imaging system 12 and a separate first image processor (not illustrated) can be provided to process or pre-process image data from the imaging system 12. The image data from the image processor can then be transmitted to the navigation processor 74. It will be understood, however, that the imaging systems need not perform any image processing and the image data can be transmitted directly to the navigation processing unit 74. Accordingly, the navigation system 10 may include or operate with a single or multiple processing centers or units that can access single or multiple memory systems based upon system design. [0050] In various embodiments, the imaging system 12 can generate image data that may be used to compose the image 78 and define an image space that can be registered to a patient space or navigation space that is defined by and/or relative to the patient 14. In various embodiments, the position of the patient 14 relative to the imaging system 12 can be determined by the navigation system 10 with the patient tracking device 56 and the imaging system tracking device(s) 22 to assist in and/or maintain registration. Accordingly, the position of the patient 14 relative to the imaging system 12 can be determined.
[0051] Manual or automatic registration of the image space to the subject space can occur. In various embodiments, the registration can occur by matching fiducial points in image data with fiducial points on the patient 14. Registration of image space to patient space allows for the generation of a translation map between the patient space and the image space. According to various embodiments, registration can occur by determining points that are substantially identical in the image space and the patient space. The identical points can include anatomical fiducial points or implanted fiducial points. Exemplary registration techniques are disclosed in U.S. Pat. App. No. 12/400,273, filed on March 9, 2009, now published as U.S. Pat. App. Pub. No. 2010/02281 17 and in U.S. Pat. No. 9,737,235, issued August 22, 2017, both incorporated herein by reference.
[0052] According to various embodiments, the imaging system 12 can be used with an un-navigated or navigated procedure. In a navigated procedure, a localizer and/or digitizer, including either or both of an optical localizer 52 and/or an electromagnetic localizer 50, 55 can be used to generate a field and/or receive and/or send a signal within a navigation domain relative to the patient 14. The navigated space or navigational domain relative to the patient 14 can be registered to the image 78. Correlation, as understood in the art, is to allow registration of a navigation space defined within the navigational domain and an image space defined by the image 78. The patient tracker or dynamic reference frame 56 can be connected to the patient 14 to allow for a dynamic registration and maintenance of registration of the patient 14 to the image 78.
[0053] Once registered, the navigation system 10 with or including the imaging system 12, can be used to perform selected procedures. Selected procedures can use the image data generated or acquired with the imaging system 12. Further, the imaging system 12 can be used to acquire image data at different times relative to a procedure. As discussed herein, image data can be acquired of the patient 14 prior to the procedure for collection of automatically registered image data or cine loop image data. Also, the imaging system 12 can be used to acquire images for confirmation of a portion of the procedure. Thus, image data may be acquired at any appropriate time and may be registered to the patient 14.
[0054] Registration and navigated procedures are discussed in U.S. Patent No. 8,238,631 , incorporated herein by reference. Upon registration and tracking of the instrument 62, a graphic representation 90 (e.g., an icon, indicium, animation or other or visual representation) may be displayed relative to, including overlaid (e.g., superimposed) on, the image 78. The image 78 may be an appropriate image and may include one or more 2D images, such as 2D images that are acquired at different planes. Images may also be a 3D image, or any appropriate image as discussed herein.
[0055] In addition to registering the subject space to the image space, however, the imaging plane of the US imaging system 12 can also be determined. By registering the image plane of the US imaging system 12, imaged portions can be located within the patient 14. For example, when the image plane is calibrated to the tracking device(s) 22 associated with the US housing 16 then a position of an imaged portion of the heart 15, or other imaged portion, can also be tracked as disclosed in U.S. Patent Nos. 6,379,302; 6,669,635; 6,968,224; 7,085,400; 7,831 ,082; 8,320,653; 8,811 ,662; and 9,138,204 all of which are incorporated herein by reference.
[0056] Once the patient 14 is in condition for a procedure, the patient 14 may be positioned on the table 40, as noted above. The display 26 and/or the display 77 may display various information regarding the patient 14 and/or other information selected by the user 18. As noted above, the display 26 may illustrate the image 78 that may be acquired with the imaging system 12. The image 78 may also be displayed on the display 77. Also as noted above, the instrument 62 may be tracked and the graphic representation 90 of the instrument 62 may be displayed on the display 26, such as relative to the image 78. In addition and/or alternatively, the graphic representation 90 may also be displayed relative to a patient avatar 120.
[0057] The patient avatar 120 may be illustrated on a selected display, as illustrated in Figs. 2A and 2B, such as the display 26. The patient avatar 120 may be based on a general avatar that includes various features of the patient 14, as noted herein, but is sized to the current and specific patient. The avatar 120 may be illustrated as a two-dimensional (2D) image and/or a three-dimensional (3D) image. The avatar 120 may be any appropriate shape to assist in identifying various positions and relative positions of portions of the patient 14. The avatar 120, according to various embodiments, is sized to the specific patient as disclosed herein.
[0058] The patient avatar 120 may generally illustrate a portion of the patient 14 and may be sized relative to the patient 14. For example, the patient avatar may illustrate an inferior portion of a neck 122, a pectoral region 124, and other appropriate regions of the patient 14. The patient avatar 120 may be displayed on the display device 26 to illustrate a general representation of the patient 14 without requiring an image acquisition of the same portion of the patient 14. Therefore, the patient avatar 120 may provide a general roadmap or indication of the position of various tracked portions, as discussed further herein. The avatar 120 may however, as discussed herein, be morphed to match a relative size of the patient 14 to provide an indication of relative anatomical portions of the patient 14 and a pose of the tracked portions, such as the instrument 62, relative thereto. [0059] With continuing reference to Fig, 2 and additional reference to Fig. 3, the avatar 120 may be generally sized to the patient 14 according to one or more appropriate procedures. The avatar 120 may be sized or morphed in real time to match relative dimensions of the patient 14. The relative positions of the anatomic portions may include a relative position of a suprasternal notch and an exterior shoulder or inferior jugular vein, or any other suitable anatomical location(s). [0060] In various embodiments, for example, the user 18 may provide or include information regarding the patient 14. As noted above, the navigation processing system 70 may include various inputs that allow the user 18 to input information. Therefore, as illustrated in Fig. 3, the user 18 may input information such as a height, a weight, and a sex of the patient 14. The input information is used by the processor module 74 to execute selected instructions to size the patient avatar 120 to the actual patient 14. The resizing of the avatar 120 may be based upon recalling of a selected or known predetermined size from a database, reconfiguring or resizing the patient avatar 120 based upon the input information, or other appropriate information. Therefore, the avatar 120 may be sized to substantially match the patient 14 without acquiring an image of the patient 14 for display on the display device 26.
[0061] In addition or alternatively thereto, the user 18 may also identify or size the avatar 120 to the patient 14 with a tracked indicator or tracked member 130. For example, the tracked indicator 130, which may be an indicator or probe usable to touch a surface of the patient 14, may be moved by the user 18 to various predetermined positions on the patient 14 as indicated on the avatar 120. The tracked indicator 130 may be tracked with any appropriate tracking system, including those discussed herein according to various embodiments such as the localizer system of U.S. Patent No. 5,983,126, hereby incorporated by reference. The user 18 may move the tracked indicator 130 to various points, such as the exterior shoulder portions 132’, 134’ as illustrated by the icons 132 and 134. Additionally, the user 18 may move the tracked indicator 130 to other portions such as a suprasternal notch 136 and a sternum 140 of the patient 14 or outer boundary positions. It is understood that the tracked indicator 130 may be moved to any appropriate position and the above-noted positions are merely exemplary. Nevertheless, based upon the tracked position of the tracked indicator 130, the avatar 120 may be sized to morph the identified points to a representation of the patient 14 for display on the display device 26. Also, the points on the avatar 120 may be registered to the patient 14 by identifying them in the patient space relative to the patient 14 and identifying them on the avatar 120.
[0062] The avatar 120 may also be morphed to actual measurements taken of the patient 14. For example, a distance between the shoulder points 132’, 134’ on that the patient 14 may be measured, such as with a tape measure or other measuring device. The measured distance may be input within the navigation processing system at 77. The avatar 120 may be sized based upon the measurements.
[0063] Regardless of the technique, the avatar 120 may be matched or morphed to substantially mimic the shape of the patient 14. The avatar 120, as discussed further herein, may then be used as an illustration of a pose of the tracked instrument, such as the instrument 62, during a procedure on the patient 14. Thus, the avatar 120 may provide a representation of a position of the tracked instrument 62 without requiring an image acquisition of the patient 14, such as with a fluoroscopic imaging system. The avatar 120, as it is morphed to the specific patient may or will be displayed as a patient specific avatar. Thus, the avatar 120 may include general information, but is displayed and registered to the patient 14 as a patient specific avatar 120 including relevant locations and relative positions of various portions of the patient 14. The specific patient avatar illustrates the relative positions of the anatomy of the patient 14 that is the current patient. For example, the position of the heart 15 relative to the suprasternal notch.
[0064] The avatar 120 may be registered to the patient 14 in any appropriate manner, including those discussed above. In various embodiments, for example, when the tracked indicator 130 is tracked to determine positions of the patient 14 for marking or changing the avatar 120. The patient 14 after the various points are determined in both the patient 14 and the avatar 120 may be registered thereto. Additionally, and/or alternatively, the patient tracker 56 may be positioned on the patient 14 at an appropriate position. The position of the patient tracker 56 may be at the suprasternal notch 136’ which relates to a point on the avatar 120, such as the suprasternal notch 136. The position of the suprasternal notch 136 in the avatar 120 may be determined or known to be the position of the patient tracker 56 to allow for registration of the avatar 120 relative to the patient 14. Also, the patient tracker 56 may maintain registration even when the patient 14 moves relative to the localizer(s). It is understood that any appropriate number of the patient trackers 56 may be utilized, including each may be used for registration and/or more than one of the patient trackers 56 may be used.
[0065] The localizers, including the EM localizer 58, may also or alternatively be fixed or positioned relative to the patient 14 in an appropriate manner. For example, the localizer 58 may be fixed relative to the table 14 such that it is immovable during the procedure. Thus, the field generated with the EM localizer 58 can be at a known position relative to the patient 14 and allow for a registration of the avatar 120 to the patient 14 based upon the known position of the EM localizer 58 relative to the patient 14. The localizer 55 may also be positioned under the patient 14, such as between the patient 14 and the bed 40. Localizer 55 may include field generating portions, such as coils, that generate a field relative to the patient 14. The flat or panel localizer 55 may be positioned at a known position relative to the patient 14, such as aligning the shoulder edges 132’, 134’ to known or predetermined portions of the panel localizer 55. As the panel localizer 55 is at a known position relative to the patient 14, and the avatar 120 is morphed to the patient and 14, the avatar 120 may be registered to the patient 14 due to the position of the panel localizer 55.
[0066] Regardless, the avatar 120 may be registered to the patient 14. Registration of the avatar 120 relative to the patient 14 allows the avatar 120 to be used to assist navigation and guiding of the instrument 62, or any appropriate instrument relative to the patient 14. As the tracked position of the instrument 62 is determined with the navigation system 10, the pose of the instrument 62 may be illustrated with a graphical representation, such as the graphic representation 90 relative to the image 78 and/or the avatar 120.
[0067] As illustrated in Figs. 2-5, the instrument’s graphic representation 90 is displayed superimposed on the avatar 120. The user 18 may view the display 26 and understand the tracked position of the instrument 62 relative to the patient 14 by viewing the graphic representation 90 of the instrument 62 in the avatar 120 on display 26. The user 18 may then understand the relative position of the instrument 62 to the patient 14 and determine that the instrument 62 is generally moving and/or is positioned in a selected or procedure planned position. The user 18, thus, may understand this by viewing the display 26 and the avatar 120 without requiring additional or any image data being acquired of the patient 14. [0068] In various embodiments, for example, image data may be acquired of the patient 14. As noted above, the imaging system 12 may acquire image data of the patient 14. The imaging system 12 may generally be a non-ionizing imaging system, such as an ultrasound imaging system. The ultrasound image may generally be collected at a single plane relative to the housing 16. The US imaging system 12 with the tracker 22 may be tracked relative to the patient 14 so that the image may also be registered to the patient 14. Thus, the image 78 may also be displayed superimposed on the avatar 120, such as the image superimposition 78i. The image may also be displayed separately from the avatar, as also illustrated in Figs. 2A and 2B. Fig. 2A illustrates the image 78 at a first orientation. Fig. 2B illustrates the image 78 at a second orientation that is about 90 relative to the first orientation. The operator of the imaging system 12 may toggle between views to select the view that they find most suitable without moving the housing 16.
[0069] As the image 78 may be registered to the patient 14, a single image acquisition may occur and the image 78 may be displayed for the procedure. Further, as the image 78 is registered to the patient the graphic representation 90 of the instrument 62 may be displayed at the tracked pose on the image 78. Thus, the US imaging system 12 may be used at one point in time to acquire the image data, but it may be used to illustrate the selected portion of the patient 14, such as the heart, for an entire procedure. A constant imaging of the patient 14 is not required to illustrate or know a position of the instrument 62 as the instrument 62 is tracked and the image 78 is registered to the patient 14 due to the US tracker [0070] In a procedure, once the avatar 120 and/or the image 78 are registered to the patient, the user 18 may perform the procedure on the patient 14 with the tracked instrument(s). The procedure may be any appropriate procedure, such as placing a pacing lead and/or a catheter to assist in placing the lead. Other appropriate procedures may also be performed.
[0071] In various embodiments, the user 18 may form an incision, such as an incision 150 on the patient 14, and move the instrument 62 into the patient 14 through the incision 150. The tracked portion 60 on the instrument 62 may be tracked so that a position of the instrument 62 may be displayed relative to the avatar 120 and/or the image 78, as noted above. The instrument 62 may then be moved relative to the patient 14. In various embodiments, for example, the instrument 62 may be a catheter and/or lead for an implant, or a stylet or a delivery system for an implant such as a leadless pacemaker or a stent. As is understood by one skilled in the art, an implant, including an electrode, may be positioned relative to one or more portions of the heart 15 of the patient 14.
[0072] With continuing reference to Fig. 5 and additional reference to Fig. 6, the position of the instrument 62 may be illustrated with the graphic representation 90 over time and/or instantaneously relative to the avatar 120. The graphic representation 90 illustrates the pose of the instrument 62 tracked relative to the patient 14. The graphic representation 90 may be viewed by the user 18 to understand a real-time position of the instrument 62 within the patient 14 via the avatar 120. Again, the avatar 120 may be generated without requiring image data of the patient 14, such as with fluoroscopy. [0073] The avatar 120 may also assist the user 18 in identifying or determining an orientation of the instrument 62 as being in a selected or in a not- selected pose relative to the patient 14. For example, as illustrated in Fig. 6, the avatar 120 may be viewed by the user 18 during a procedure. The user 18 may view the representation 90 of the instrument 62 as it moves in a selected direction and/or position. The tracked pose of the instrument 62 may be updated in substantially real time and displayed relative to the avatar 120. Therefore, the graphic representation 90 of the instrument 62 when it is moving in a not-selected position or and/or pose may also be illustrated and understood by the user 18. As illustrated in Fig. 6 a representation 90x may illustrate movement and/or pose toward a neck 122 of the patient 14. This may be a not-selected pose when the instrument 62 is selected to be moving toward the heart 15 of the patient 14. Therefore, the user 18 may view the graphic representation 90x and understand that the instrument 62 is moving away from the heart 15 of the patient 14. A representation of the heart 15 may also be displayed on the avatar 120 including the image 78 and/or with a heart representation 15i that is included in the avatar 120 and appropriately located with the patient specific information.
[0074] Thus, the user 18 may view and understand the position of the instrument 62 during the placement of an implant, which may include, for example, a stimulation lead for a selected system. The lead may include a pacing electrode be for a cardiac stimulation system, such as the AZURE®, ADAPTA®, or other appropriate heart pacemaker or defibrillation systems sold by Medtronic, Inc. having a place for business in Minnesota, USA. [0075] Turning reference to Fig. 7, a method 200 for determining and displaying a patient specific avatar is illustrated. The process may begin in Block 210 and include recalling a standard avatar in Block 214. In recalling a standard avatar the processor, such as the processor 74, may recall a standard avatar including various anatomical features, including exterior dimensions of a patient's anatomy and selected internal anatomy such as a heart, physical location of the heart, jugular vein, or the like.
[0076] Information regarding the patient, such as the specific patient, may be received in Block 218. The information regarding the patient may be any appropriate information, including that discussed above. Therefore, the information may include general dimensions of the current patient, including a height, weight, sex, or the like. Information received in Block 218 may also include specific tracked points of predetermined portions (e.g., shoulder edges) of the patient 14. Further the received information may include measurements of the patient. Regardless, the received information of the patient may be patient specific information regarding the patient for a current or real-time procedure.
[0077] A patient specific avatar is then generated in Block 230. The generation of a patient specific avatar may include morphing or fitting the standard avatar based upon the received information. In morphing the standard avatar, the method may include altering or scaling dimensions of the standard avatar based upon a ratio given the information received regarding the current patient. The ratio may be based upon the known dimensions of the current patient based upon known or average dimensions of various anatomical portions relative to one another based upon the dimensions of the specific patient. This may also include having predetermined average positions and relative positions of various anatomical features. In addition and/or alternatively thereto, the tracked position of the instrument may identify specific anatomical locations. The standard avatar may then be morphed based upon the tracked specific locations for the specific patient. Thus, the avatar may be morphed based upon selected instructions including an algorithm to generate the patient specific avatar in Block 230.
[0078] After generating the patient specific avatar in Block 230, the patient specific avatar may be displayed in Block 234. As discussed above, the display 26 may display the patient specific avatar for selected procedures. The display of the patient specific avatar may allow the user 18 to identify and understand relative positions or locations of various portions of the current patient 14.
[0079] An instrument may then be tracked, optionally and according to various embodiments, in Block 238. Although the tracking of the instrument is an optional part of the method 200, such tracking may allow for information regarding a pose of an instrument to be determined. Regardless of whether the instrument is tracked, the method 200 may include receiving tracking information in Block 240.
[0080] Receiving tracking information in Block 240 may include tracking one or more tracking devices to determine a pose of various elements or portions, such as the pose of the instrument 62 via the instrument tracker 60 and/or the patient 14 via the patient tracker 56. The patient specific avatar may be registered to the patient based upon the known location of one or more specific portion(s) of the patient, such as the positioning of the patient tracker 56 in the suprasternal notch of the patient 14. The suprasternal notch of the avatar 120 may then be registered for determination of the pose of the instrument 62 relative to the patient tracker 56. Therefore, the pose of the instrument 62 may be displayed on the display device 26 such as the superimposed graphic representation 90 on the avatar 120.
[0081 ] Thus, the pose of the instrument may be superimposed on the avatar 120 in Block 244. Again, this allows a user 18 to view and understand a pose of the tracked instrument 62 on a representation of the patient 14 without requiring image data to be acquired to the patient 14. While image data of the patient 14 may be acquired, such as with the US imaging system 12, such image information is not required. The image data information from the US imaging system 12 may assist in finalizing or determining a specific location of the instrument 62 relative to a selected portion, such as a specific portion of the heart 15 of the patient 14. According, only a selected portion or area of the patient may be imaged. However, the position of the instrument 62 may be determined without the image data. Thus, the user may understand a position of the instrument 62, including a movement of the instrument 62 toward a selected position, such as the RV septum of the patient 14.
[0082] The method 200 may then end in Block 250. And ending the method 250 it is understood that the method 200 may be repeated as often as necessary or selected to allow for up to date and/or real-time tracking and illustration of a pose of the instrument 62 relative to the avatar 120. Thus, the method 200 may be repeated at a selected rate and for a selected period, such as selected by the user 18. [0083] As noted above, moving the instrument 62 relative to the patient 14, such as relative to the heart 15, may be accomplished while tracking a pose of the instrument 62. As noted above, tracking a pose of the instrument 62 can involve attempting to determine a physical location and/or orientation of the instrument 62 in the physical space. The tracked posed may include a location which includes X, Y, and Z axis positions and orientation which may include yaw, pitch, and roll orientations in space. The physical space may also be referred to as navigation space and is generally defined relative to the patient 14 and/or the heart 15. The patient tracker 56 allows the pose of the patient 14, including the heart 15, to be tracked. Tracking a pose relative to a portion of the patient 14, including the heart 15, may include systems such as those disclosed in U.S. Pat. No. 7,697,972 issued April 13, 2010; U.S. Pat. No. 8,046,052 issued October 25, 201 1 ; U.S. Pat. No. 8,180,428 issued May 15, 2012; and U.S. Pat. No. 8,663,120 issued March 4, 2014, all of which are herein incorporated by reference.
[0084] With continuing reference to Figs. 1 - 7, and additional reference to Fig. 8, the instrument 62 may be tracked relative to the patient 14 which may be displayed such as within the graphic representation 90 relative to selected portions including the avatar 120, the image 78, and/or an atlas 78m, as illustrated in Fig. 8. The atlas 78m may be based upon the selected information, such as based on a plurality of images of patients that have been averaged and portions thereof identified. For example, the atlas 78m may be a model based on the image data from a plurality of images. The atlas 78m (also referred to as a model) may have selected portions therein that are identified. In various embodiments, the atlas 78m may include the following identified locations: a right ventricle 300, a superior vena cava ostium 304, a tricuspid valve 308, a right ventricle 312, and a right ventricle septum 316.
[0085] With reference to Fig. 8, the image 78 is an example of one image in a set of images that can be used to build the atlas 78m. The atlas 78m may also be generated using the patient’s pre-procedure imaging like a CT or MRI. The image 78 may be obtained with a selected imaging system, including the US imaging system 12, as discussed above. A patient specific anatomy model may be generated using multiple 2D/3D views/planes stitched together, for example to reconstruct the endocardium of the heart. To facilitate building the atlas 78m, a clinician may identify various relevant portions or locations within the image 78. The user 18 may identify portion(s) or location(s) in the image 78 shown in the right-hand side of Fig. 8 as the superior vena cava entrance 304’, the tricuspid valve 308’, and/or the right ventricle 312'. Other locations may include endocardium or blood pool boundary may include the identification of the portions/locations within the heart. The location identification may be done manually by the user 18 and/or with assistance of an automatic location identification, such as segmenting a valve, etc.
[0086] Multiple images similar to the image 78, e.g., multiple similar images from a population of patients other than the patient 14, with similar identified portion(s) or location(s) may be combined to build the atlas 78m to illustrate the identified portion(s) or location(s) on a model of the imaged anatomy or heart, which model represents a typical or average patient anatomy. Alternatively, the atlas 78m may be built only from images of the anatomy of the patient 14, and the portion(s) or location(s) of interest applied to the atlas based on identification within the images of the patient’s own anatomy, and/or based on a compilation of images from other patients’ anatomy and an estimated and/or average position of the portion(s) or location(s) of interest within the compilation of images. The atlas 78m may also be morphed to actual measurements taken of the patient 14.
[0087] Based upon the user's 18 identification in the image 78 of the various portions/locations, which may include all of or selected number of portions/locations, a processing module or unit, such as the processor 74, may execute selected instructions to register or identify various portions in the image 78 and/or on the atlas 78m for viewing by the user 18. For example, in the atlas 78m the right ventricle septum may be identified as a selected point including a target 330. The image 78 may also, therefore, have the target identified thereon 330x to illustrate a proposed or selected position of the right ventricle septum (RV septum), e.g., to serve as a target for delivery of a catheter, stylet and/or pacing lead, or a leadless pacemaker. The predetermined portions in the model 78m, therefore, may be used to assist in identifying portions of the image 78 based on one or more portions identified by the user 18 and/or appropriate image identification systems. Therefore, the model 78m may be used to assist in identifying portions within the image 78 to assist in the procedure. The user may view the image 78 and identify or evaluate the target 330x prior to and/or during the procedure.
[0088] The image 78, as discussed above, may also include a plurality of images that are generated with the imaging system 12. The US imaging system
12 may acquire images of the patient 14, including the heart 15, in a plane. As the US housing 16 is moved, the plane of the image relative to the patient 14 may also change. Thus, the image 78 may include a plurality of images that are illustrated in one or more planes. For example, with reference to Fig. 9, schematic images that may represent an approximately coronal to the body co-ordinate system image 78c may be acquired and displayed and/or an image that is approximately axial to the body co-ordinate system 78ax may be displayed. The two images 78ap and 78ml may both be displayed and registered to the patient 14, as discussed above. Further, the tracked position of the instrument 62 may be displayed relative to one or both of the images 78ml and 78ap. Also, both of the images 78ml and 78ap may assist in registering or relating of the images to the model 78m. Additionally, the user 18 may view the images, as discussed above, and identify portions or locations thereon and of the model 78m that may be used to assist in identifying various other portions. For example, the user 18 may identify the tricuspid valve 308’ and the superior vena cava entrance 304’ and the model 78m may be used to identify various other portions or locations, such as the target 330x on the RV septum 330.
[0089] In addition to providing more than one view of the heart 15 of the patient 14, the two images 78ap and 78ml may also assist in registration to the model 78m. Thus, the identification of anatomy in the images 78 and/or one or both of the images and 78ap and 78ml, may be assisted by the registration of more than one image relative to the model or atlas 78m. As discussed above, the ultrasound imaging system 12 is registered to the patient 14, thus the images acquired with of the imaging system 12 may also be registered to the patient 14. The registration of the images to the patient may be based upon the tracking of the imaging system 12, such as with the imaging system tracking device 22 and the tracking of the patient 14, such as with the patient tracking device 56. Once the image 78 is registered to the patient 14, the instrument 62 may be tracked relative to the patient 14 and its pose may be displayed relative to the image 78. As discussed above, the pose of the instrument 62 may also be displayed relative to the avatar 120. The graphic representation 90 allows the user 18 to view the pose of the instrument 62 relative to both of the image 78 and the avatar 120.
[0090] Further, the instrument 62 may be tracked in real time. Therefore, the instrument may be displayed relative to the avatar 120 and/or the image 78 to provide a real time graphical representation of the instrument 62 relative to the heart 15, based upon a graphical representation displayed relative to the image 78. In addition to viewing the tracked pose of the instrument 62, the graphical representation may show the pose of the instrument 62 over time. Therefore, the graphic representation 90 may illustrate or provide a sense of an entire path of the instrument 62 as it moves from the incision 150 in the patient 14 to the current location within the patient 14. Further, the graphic representation 90 may provide a real-time feedback to the user 18 of a movement of the instrument 62.
[0091] In various embodiments, the images 78 may be acquired or based upon image data acquired at a selected time. For example, prior to the incision 150, image data may be acquired with the imaging system 12 to generate the image 78. Image 78 may thus be registered to the patient 14 and displayed with a display device 26 either alone and/or superimposed on the avatar 120. Nevertheless, the image 78 may not be a real time or current image and may be a static image. The instrument 62, however, may be tracked in real time and provide a current determination of the pose of the instrument 62. The graphic representation 90 may, therefore, be updated in real time and at a selected rate to illustrate a current position of the instrument 62 and/or a current shape or condition of the patient anatomy. The display of the graphical representation may also be updated in a selected rate to provide a perception of motion of the instrument 62. For example, a refresh of the pose of the instrument 62 and of the graphic representation 90 may be 10 times a second, 30 times a second, or any appropriate rate such as about one time a second to about 120 times per second. Thus, the pose of the instrument 62 may be illustrated to show a motion within the display device 26.
[0092] With reference to Fig. 10A and 10B, the image 78 may be displayed on the display device 26. The image 78 may not change over time as the image 78 may be a static image that is acquired at a selected time, such as prior to positioning the instrument 62 within the patient 14. Nevertheless, the tracking device 60 may be tracked in real time on the device 62. Thus, as illustrated in Fig. 10A, the graphic representation 90 may include a first or initial time graphic representation 90t1 . The graphic representation 90t1 may represent a first graphic representation at a selected time, that may be arbitrarily referred to as a first or initial time. After a selected period of time, a second graphic representation 90t2 may be displayed on the display device 26, as illustrated in Fig. 10B.
[0093] The graphic representation 90t2 may be displayed or have a selected position that is displaced a certain distance of 350 relative to the first graphic representation 90t1 . The distance 350 may be any appropriate distance and may be based upon the various physical features, such as movement of a beating heart. The heart 15 of the patient 14 may beat (i.e., move due to the heartbeat), as is understood by one skilled in the art. Therefore, the graphic representation 90t2, including a terminal end 90x, may be shown in different positions based upon movements of the instrument 62. For example, as one skilled in the art will understand, when the terminal end of the instrument 62 contacts the RV septum 316, the terminal end of the instrument 62 may move due to the beating of the heart 15. Therefore, even if the user 18 is not moving the instrument 62, the instrument 62 may appear to move in real time. Real-time tracking of the instrument 62 with the tracking device and 60 may allow for a representation of the movement of the instrument 62 to be displayed on the display device 26. As discussed above, the movement of the instrument may be updated and displayed at any appropriate rate to provide a perception of movement to the user 18. Thus, the user 18 may be provided with additional information regarding a position of the instrument 62, such as being in contact with of the RV septum 316.
[0094] With reference to Fig. 11 , a method or process 370 is illustrated. The process 370 may be a process that is executed by one or more of the processing units, such as processing unit 74. The method 370 may begin at start Block 374. After starting the process in Block 374, the process 370 may determine a first position or pose of the instrument in Block 378. Determining the first pose of the instrument may be performed with the navigation system, as discussed above. The pose of the instrument 62 may be known relative to the patient 14, or portions thereof such as the heart 15, in light of the tracking of the patient with the patient tracker 58. [0095] After determining a pose of the instrument at a first time, an output for display of the determined pose of the instrument may be made in Block 382. The output of the determined pose may be based upon the registration of the image 78 and/or the avatar 120 to the physical space defined by the patient 14, as discussed above. Thus, the output may display the graphic representation 90 at a first position, for example of the first pose 90t1 , as illustrated in Fig. 10A.
[0096] After determining and outputting the first pose, a wait or pause for a selected time may occur in Block 384. The wait for a selected time may be any appropriate time, such as about one second to about 120 seconds. The wait time may be any appropriate time to assist in illustrating a motion to the user 18, as discussed above. After the wait time, the second pose may be determined in Block 388. The second pose may be determined in the same manner as the first pose was determined in Block 384. An output for display of the second pose may be made in Block 392 to display the graphical representation of the instrument 62 at a second time, such as the graphic representation 90t2. As noted above, therefore, the pose of the instrument 62 may be displayed over a series of the events and/or time to show movement to the user 18.
[0097] A determination can be made to continue tracking at Block 394. If tracking is to be continued, a YES path 398 may be followed to again determine a first or subsequent pose in Block 378. The process 370 may provide an illustration to the user 18 for an appropriate period of time. If a tracking is not to be continued, a NO path 400 may be followed to end the process in Block 404. The process 370 may be used to illustrate and determine a pose of the instrument 62 over time for illustration relative to the image 78, as illustrated in Fig. 10A and 10B. [0098] The display 26 may display the image of the patient 14 that is acquired with the imaging system 12. As discussed above, the image 78 may be an image that is acquired in a plane from the ultrasound housing 16. Therefore, rotation of the image may change a perspective of the image 78. With reference to Fig. 12A, the image may be viewed over the plane 78f or from an edge 78e. The edge image 78e may allow for a viewing and understanding of a perspective of the instrument 62, as illustrated by the graphic representation 90, relative to a plane of the image. In various embodiments, the image acquired with the imaging system 12 may be a static image that is acquired with the imaging system 12. It is understood, however, that a live or current image may also be acquired and displayed. Nevertheless, the position of the instrument 62 relative to the plane of the image may be illustrated by rotating the image, such as showing it from an edge as illustrated in the edge image 78e.
[0099] Further, the distance of the instrument 62 that is within the plane of the image may also be illustrated. For example, a first and second distance icon 410, 412 may be included or illustrated superimposed on both of the images 78e, 78f. The distance icons may be used to illustrate the distance that the instrument 62 is in the plane of the image. Thus, the user 18 may also understand that the portion of the instrument that is within the plane of the image and that other portions are not within the plane of the image.
[0100] The distance icon 410, 412 may be sized depending upon the length or portion of the instrument 62 that is within the plane of the image 78. Therefore, the first distance icon 410 may have a first distance or size 420. The second distance icon 412 may have a second distance or size 424. The two dimensions 420, 424 may be different or the same depending upon the amount of the instrument or length of the instrument that is within the plane of the image 78. If the instrument has a greater length within the plane, the distance icons 410, 412 may have greater dimensions. If any dimension of the instrument is smaller in the plane, then the dimension icons may be smaller. Regardless, the dimension icons 410, 412 may be displayed, such as superimposed on, the image 78 to represent the distance of the instrument within the image plane. This may allow the user 18 to understand or determine the distance of the instrument 62 within a plane of the imaging system and/or assist in moving of the imaging system to acquire a different perspective than in the image 78 relative to the instrument 62.
[0101] According to various embodiments, the display 26 may display the image of the patient 14 that is acquired with the imaging system 12 at more than one location. In other words, more than one image may be displayed for each position of the imaging system 12 at which image data is acquired. Moreover, one or more image data portions may be merged and/or used in a reconstruction of a displayed image. As discussed above and herein, all of the images and/or the reconstructed image may be registered to the patient 14 due at least to the tracking of the imaging system 12.
[0102] With reference to Fig. 12B, for example, the image 78 may be an image that is acquired in a plane from the ultrasound housing 16. Therefore, rotation of the image may change a perspective of the image 78. The image may be viewed as acquired at a first location (e.g., inferior from the suprasternal notch) as over a plane 78f ’ or from an edge 78e’. These two views may be able to provide information to the user, as discussed above including location and/or dimension that the instrument graphic representations 90 passes through the plane 410’, 412’, 420’, 424’. Further, the image may be viewed as acquired at a second location (e.g., between selected ribs) as over a plane 78f” or from an edge 78e”. These two views may be able to also provide information to the user, as discussed above including location and/or dimension that the instrument graphic representations 90 passes through the plane 410”, 412”, 420”, 424”. Given that the image portions 78e’, 78f’ are acquired at a different position than the image portions 78e”, 78f”, the graphic representations 90 is illustrated to pass through the planes at the different positions and/or dimensions 41 O’, 412’, 420’, 424’, 410”, 412”, 420”, 424” and this information is displayed for the user. Further, it is understood that any appropriate number of location images may be acquired and/or displayed and two is merely exemplary.
[0103] Again, the edge image may allow for a viewing and understanding of a perspective of the instrument 62, as illustrated by the graphic representation 90, relative to a plane of the image. In various embodiments, the image acquired with the imaging system 12 may be a static image that is acquired with the imaging system 12. It is understood, however, that a live or current image may also be acquired and displayed. Nevertheless, the position of the instrument 62 relative to the plane of the image may be illustrated by rotating the image, such as showing it from an edge as illustrated in the edge image.
[0104] The instrument 62 may be tracked with any appropriate tracking system. As noted above, tracking and/or navigation systems include those disclosed above and those disclosed in U.S. Pat. No. 7,697,972 issued April 13,
2010; U.S. Pat. No. 8,046,052 issued October 25, 201 1 ; U.S. Pat. No. 8,180,428 issued May 15, 2012; and U.S. Pat. No. 8,663,120 issued March 4, 2014, all of which are herein incorporated by reference. The navigation system may allow for a determination of a pose of the instrument 62 in a physical location and/or orientation of the instrument 62 in the physical space. The tracked pose may include a location which includes X, Y, and Z axis positions and/or orientation which may include yaw, pitch, and roll orientations in space. A tracking and navigation system may also include one or more permanent magnets and/or magnetometers.
[0105] With reference to Fig. 13, the navigation processing system 70 is further configured to generate the display 26 of Fig. 13. The display 26 shows a probe housing icon 16' of the housing 16 of the imaging system 12. That is, the icon 16’ is displayed relative to, such as superimposed on, the avatar 120. This allows the user to understand the pose of the probe 16 relative to the subject and the image data acquired with the probe 16.
[0106] Shown emanating from the probe housing icon 16' is a field of view 24 of the imaging system 12. The portion of the anatomy within the field of view 24 captured by the imaging system 12 is shown as the ultrasound image 78. In the example illustrated, a heart image or model 110 is within the field of view 24. The heart image would be an actual ultrasound image captured by the imaging system 12. A 3D model would be a model created from segmented ultrasound images or from a registered CT or MRI scan. The location of the field of view 24 on the display 26 is determined based on the pose of the housing 16, as determined by tracking the position of the tracking device 22. Showing the field of view 24 on the display 26 assists the operator of the imaging system 12 with positioning the housing 16 to capture a desired portion of the anatomy. The user can see on the avatar 120 the portion of the anatomy within the field of view 24, which is displayed as the US image 78. If an area of interest is not within the field of view, the user can use the avatar 120 as a guide for maneuvering the housing 16 until the area of interest is within the field of view 24 and displayed in the ultrasound image 78. For example, the US image 78 of Fig. 13 is captured when the probe 16 is in the orientation of the probe housing icon 16' shown in Fig. 13. If the area of interest is not visible in the US image 78 of Fig. 13, the user can use the avatar 120, the heart image or model 1 10, the housing icon 16', and the field of view 24 as guides for maneuvering the actual housing 16 to bring the area of interest into the field of view 24. More specifically, if the area of interest is at the mid-papillary level of the heart, for example, the user can maneuver the actual housing 16 until the field of view 24 is at the mid-papillary level of the heart image or model 1 10. This is the case in Fig. 14, where the housing 16 has been maneuvered such that the field of view 24B extends through the mid-papillary level of the heart image or model 1 10, thereby resulting in the capture of US image 78B at the mid-papillary level of the heart. Thus, the user or selected processor executing instructions may view and understand the a current position of the probe and move the probe to a new position based on the current position.
[0107] The navigation processing system 70 can also be configured to generate the display 26 of Fig. 14. In the example illustrated, the display 26 includes three panels. The center panel shows the avatar 120 displaying the subject’s heart, or other relevant anatomy, as the heart image or model 1 10. Also shown on the center panel is the probe housing icon 16', a first field of view 24A, and a second field of view 24B. The first field of view 24A corresponds to the US image 78A on the left of the display 26. The second field of view 24B corresponds to the US image 78B on the right side of the display 26. The first field of view 24A and the second field of view 24B may be captured by two different housings 16, or by the same housing 16 at different times.
[0108] With renewed reference to Fig. 8 and additional reference to FIG. 15, the atlas model 78M may be generated by capturing a plurality of US images of the subject using the imaging system 12. The images collected of the subject, such as the heart, may also be used to morph an atlas of the heart of the current subject. In various embodiments, a user must thus maneuver the housing 16 to capture an entirety of the anatomy, such as an entirety of the heart. It is understood, however, that collecting a sufficient amount of image data of the subject may be appropriate to morph an atlas or generate an atlas.
[0109] For some users, it may be difficult to identify areas of the heart that have been captured and areas of the heart that have yet to be captured, whether to ensure a capture of an entirety of the heart or of all of the areas necessary to generate the appropriate atlas or model. The navigation processing system 70 may be configured to generate the display 26 of FIG. 15, which includes four panels displaying different sides of the heart image or model 1 10. The heart image or model 1 10 is displayed in any suitable manner that distinguishes areas of the heart that have been scanned by the imaging system 12 versus areas of the heart that have not been scanned by the imaging system 12. During the scanning process, the navigation system 70 tracks movement of the housing 16 relative to the anatomy. In the example illustrated, first portions 1 12 of the heart image or model 1 10, which are illustrated with solid lines, have already been imaged by the imaging system 12. The navigation processing system 70 tracks the position of the housing 16 relative to the anatomy, in this case the heart, to identify Second portions 114 of the heart image or model 1 10, which are illustrated in phantom, have yet to be imaged. The navigation processing system 70 is configured to update the display 26 as the heart is scanned to reflect the additional scanning. For example, and with respect to the images of Fig. 15, as the atriums of the heart are scanned the navigation processing system 70 is configured to update the heart image or model 110 to convert the atriums from second portions 1 14 in phantom to first portions 1 12.
[0110] Thus, during the scanning process, the navigation system 70 tracks movement of the housing 16 relative to the heart, or other portions of the anatomy being scanned, to identify portions of the heart that have been scanned versus portions that have not been scanned. The navigation system 70 continuously updates the images of Fig. 15 as described above to distinguish the portions of the heart that have been scanned versus the portions that have yet to be scanned based on tracked movement of the housing 16 relative to the heart.
[0111] The imaging system 12, including the ultrasound probe, may be moved relative to the subject to acquire an appropriate amount of image data, as discussed above. As illustrated in Fig. 15, the imaging and/or navigation processing system may be used to identify portions of the heart that have not been imaged or not imaged enough to generate a selected feature or type of image for display. For example, as discussed above, the image data acquired may be used to generate an image for analysis of the subject. A three-dimensional or volumetric image of a portion of the subject, such as the heart, may be generated based upon the image data acquired with of the imaging system 12. For example, the ultrasound probe generates data in a plane. Planes may intersect each other when generated or collected with the ultrasound probe. As the ultrasound probe is tracked during acquisition of the image data planes that may move and change relative to one another, such as being orthogonal. The image data acquired at different perspectives (e.g., different planes) that are known relative to one another allows a reconstruction to be made. The reconstruction may be of the heart as the image 110. The reconstruction may be a three-dimensional reconstruction based upon the image data collected with the ultrasound probe and the processor may illustrate portions of the heart needed to be imaged to reconstruct the 3D image, as noted above. Reconstruction techniques may include, but are not limited to, the following: computed tomography, specifically ultrasound computed tomography; and NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis, Ben Mildenhall, Pratul P. Srinivasan, Matthew Tancik, Jonathan T. Barron, Ravi Ramamoorthi, Ren Ng (August 3, 2020)
(https://arxiv.org/abs/2003.08934), which is incorporated herein by reference in its entirety.
[0112] With reference to Fig. 16, the navigation processing system 70 is configured to generate a display 26 including a sensor icon 160 arranged relative to the avatar 120. The position of the sensor icon 160 represents the position of any of the localizers relative to the subject 14, such as any of the EM localizers 50, 55, or 58. Thus, a user may understand by viewing the avatar 120 the pose of the EM localizers 50, 55, or 58. [0113] As discussed above and understood by one skilled in the art, the EM localizers 50, 55, or 58 may emit a field that induces a current in a tracking device, such as the tracking device 22. The display 26 also includes a range icon 162, which identifies the outer boundaries of the range of the particular localizer displayed. The range icon 162 identifies that extent (e.g., area or volume), in which the field emitted by the EM localizer is effective for navigation. Similarly, a field of view of an optical localizer may be displayed as the range icon 162. The range icon may be displayed relative to the subject aviator 120.
[0114] Displaying the sensor icon 160 and the range icon 162 facilitates positioning of the subject 14 and/or the localizer 50, 55, or 58 to capture an area of the body of interest. That is, the range icon 162 illustrates that extent to which the EM localizer 50, 55, or 58 is effective. As the avatar 120 illustrates the subject to be imaged and/or have tracking devices tracked relative thereto, the range illustrates the area in which a tracking device may be effectively tracked. Thus, the user may view the range icon 12 and decide if the volume included is effective for the procedure or not.
[0115] The localizer is registered in any suitable manner. For example, registration may be carried out with a reference tracker on the subject’s body at a predefined location, such as the reference tracker 56. That is, the localizers 50, 55, or 58 may sense a reference marker placed on the subject at a known anatomical location and the localizer may be used to determine the pose of the localizer relative thereto. Further, the localizer range 162 may be relative to a second localizer that is tracked with a first localizer. [0116] Fig. 17 illustrates another exemplary display 26 in accordance with the present disclosure. The display 26 of FIG. 17 displays the avatar 120 including a first respiratory sensor icon 170 and a second respiratory icon 172 relative to the avatar 120. The first respiratory icon 170 represents the position on the subject’s chest of a first respiratory sensor including an EM tracker. The second respiratory icon 172 represents the position on the subject’s back of a second respiratory sensor including an EM tracker. The trackers may be placed in the same general vicinity by convention. One region of high motion (e.g., substernal/subxiphoid region) and one region of lesser motion (upper back). Or alternatively the patch may be placed and the user selects from a generic menu of where it was placed from an array of options that would make sense (regions of the upper torso). It is typically desirable to arrange the first and the second respiratory sensors so that they are aligned with respect to the orientation of the avatar 120 in FIG. 17. The relative positions of the first and the second respiratory sensor icons 170, 172 in Fig. 17 allows medical personnel to determine whether the actual respiratory sensors are aligned. The display 26 also includes a respiration meter 180, which includes additional first respiratory icons 170' and second respiratory icons 172'. During respiration, the icons 170', 172' move relative to each other to visually illustrate respiration. This allows monitoring for the phase and depth of respiration, identifying abnormalities with monitoring (e.g., tracker becomes disconnected or isn’t well attached), and/or abnormalities in the regularity of the breathing of the patient, which may result in inconsistent tracking or imaging. [0117] Example embodiments are provided so that this disclosure will be thorough and will fully convey the scope to those who are skilled in the art. Numerous specific details are set forth such as examples of specific components, devices, and methods, to provide a thorough understanding of embodiments of the present disclosure. It will be apparent to those skilled in the art that specific details need not be employed, that example embodiments may be embodied in many different forms and that neither should be construed to limit the scope of the disclosure. In some example embodiments, well-known processes, well-known device structures, and well-known technologies are not described in detail.
[0118] The foregoing description of the embodiments has been provided for purposes of illustration and description. It is not intended to be exhaustive or to limit the invention. Individual elements or features of a particular embodiment are generally not limited to that particular embodiment, but, where applicable, are interchangeable and can be used in a selected embodiment, even if not specifically shown or described. The same may also be varied in many ways. Such variations are not to be regarded as a departure from the invention, and all such modifications are intended to be included within the scope of the invention.
[0119] It should be understood that various aspects disclosed herein may be combined in different combinations than the combinations specifically presented in the description and accompanying drawings. It should also be understood that, depending on the example, certain acts or events of any of the processes or methods described herein may be performed in a different sequence, may be added, merged, or left out altogether (e.g., all described acts or events may not be necessary to carry out the techniques). In addition, while certain aspects of this disclosure are described as being performed by a single module or unit for purposes of clarity, it should be understood that the techniques of this disclosure may be performed by a combination of units or modules associated with, for example, a medical device.
[0120] In one or more examples, the described techniques may be implemented in hardware, software, firmware, or any combination thereof. If implemented in software, the functions may be stored as one or more instructions or code on a computer-readable medium and executed by a hardware-based processing unit that may also be referred to as a processor. Computer-readable media may include non-transitory computer-readable media, which corresponds to a tangible medium such as data storage media (e.g., RAM, ROM, EEPROM, flash memory, or any other medium that can be used to store desired program code in the form of instructions or data structures and that can be accessed by a computer).
[0121] Instructions may be executed by one or more processors or processor modules, such as one or more digital signal processors (DSPs), general purpose microprocessors, application specific integrated circuits (ASICs), field programmable logic arrays (FPGAs), or other equivalent integrated or discrete logic circuitry. Accordingly, the term “processor” or “processor module” as used herein may refer to any of the foregoing structure or any other physical structure suitable for implementation of the described techniques. Also, the techniques could be fully implemented in one or more circuits or logic elements.
[0122] The following examples are a non-limiting list of clauses in accordance with one or more techniques in this disclosure. [0123] Example 1. A system configured to assist a user with navigating an instrument relative to a subject, the system comprising: an imaging system configured to image an anatomy, the imaging system including a probe maneuverable by the user; a tracking system including a localizer, a probe tracker configured to be mounted to the probe, and a subject tracker configured to be mounted to the subject; a display device including a display screen; and a processor configured to: generate an avatar of the subject and display the avatar in a first area of the display screen; display an anatomy icon of the anatomy on the avatar in the first area of the display screen; display a probe icon and a field of view icon of the probe in the first area of the display screen at a position relative to the avatar corresponding to position of the probe relative to the anatomy; and display a first image of the anatomy captured by the probe in a second area of the display screen, the anatomy in the first image corresponding to the anatomy icon within the field of view.
[0124] Example 2. The system of example 1 , wherein the imaging system is an ultrasound imaging system, and the probe is an ultrasound probe.
[0125] Example s. The system of example 1 , wherein the anatomy includes a heart.
[0126] Example 4. The system of example 1 , wherein the processor is configured to display the first image at a first orientation or a second orientation that is different than the first orientation without the probe being moved.
[0127] Example s. The system of example 1 , wherein the processor is further configured to: display a second image of the anatomy in a third area of the display screen, the second image shows a different area of the anatomy than the first image.
[0128] Example 6. The system of example 5, wherein the second image is captured by the probe before or after the probe captures the first image; and wherein the processor is configured to display another field of view icon on the display corresponding to the second image.
[0129] Example 7. A system configured to assist a user with navigating an instrument relative to a subject, the system comprising: an imaging system configured to image an anatomy, the imaging system including a probe maneuverable by the user; a tracking system including a localizer, a probe tracker configured to be mounted to the probe, and a subject tracker configured to be mounted to the subject; a display device including a display screen; and a processor configured to: display a first anatomy icon of the anatomy in a first area of the display screen; and distinguish visually on the first anatomy icon a first portion of the anatomy that has been imaged by the probe and a second portion of the anatomy that has not been imaged by the probe.
[0130] Example s. The system of example 7, wherein the processor is further configured to display a second anatomy icon of the anatomy in a second area of the display screen, the second anatomy icon shows a different angle of the anatomy as compared to the first anatomy icon.
[0131] Example 9. The system of example 8, wherein the processor is further configured to display a third anatomy icon of the anatomy in a third area of the display screen, and display a fourth anatomy icon of the anatomy in a fourth area of the display screen; wherein each of the first anatomy icon, the second anatomy icon, the third anatomy icon, and the fourth anatomy icon are orientated at different angles; and wherein the processor is configured to distinguish visually on each of the second anatomy icon, the third anatomy icon, and the fourth anatomy icon the first portion of the anatomy that has been imaged by the probe and the second portion of the anatomy that has not been imaged by the probe.
[0132] Example 10. The system of example 7, wherein the anatomy is a heart.
[0133] Example 1 1. A system configured to assist a user with navigating an instrument relative to a subject, the system comprising: a tracking system including a localizer and a subject tracker configured to be mounted to the subject; a display device including a display screen; and a processor configured to: generate an avatar of the subject and display the avatar on the display screen; display an anatomy icon of an anatomy on the avatar; and display a localizer icon and a range icon of the localizer on the display screen at a position relative to the avatar corresponding to position of the localizer relative to the subject.
[0134] Example 12. The system of example 1 1 , wherein the localizer is an electromagnetic localizer.
[0135] Example 13. The system of example 12, wherein the electromagnetic localizer is suspended over an anterior side of the subject.’ [0136] Example 14. The system of example 12, wherein the electromagnetic localizer is at a posterior side of the subject.
[0137] Example 15. The system of example 12, wherein the electromagnetic localizer is mounted at a side of the subject. [0138] Example 16. A system configured to assist a user with navigating an instrument relative to a subject, the system comprising: a respiratory monitoring system including a first respiration sensor and a second respiration sensor; a display device including a display screen; and a processor configured to: generate an avatar of the subject and display the avatar on the display screen; display a first respiration sensor icon on the avatar at a first location corresponding to placement of the first respiration sensor on the subject; and display a second respiration sensor icon on the avatar corresponding to placement of the second respiration sensor on the subject.
[0139] Example 17. The system of example 16, wherein the first respiration sensor and the second respiration sensor include electromagnetic localizers.
[0140] Example 18. The system of example 16, wherein the processor is further configured to display on the display screen a respiration meter including a first meter icon and a second meter icon.
[0141] Example 19. The system of example 18, wherein the processor is configured to move the first meter and the second meter icon on the display screen in a manner corresponding to movement of the first respiration sensor and the second respiration sensor during respiration.

Claims

CLAIMS What is claimed is:
1 . A system configured to assist a user with navigating an instrument relative to a subject, the system comprising: an imaging system configured to image an anatomy, the imaging system including a probe maneuverable by the user; a tracking system including a localizer, a probe tracker configured to be mounted to the probe, and a subject tracker configured to be mounted to the subject; a display device including a display screen; and a processor configured to: generate an avatar of the subject and display the avatar in a first area of the display screen; display an anatomy icon of the anatomy on the avatar in the first area of the display screen; display a probe icon and a field of view icon of the probe in the first area of the display screen at a position relative to the avatar corresponding to position of the probe relative to the anatomy; and display a first image of the anatomy captured by the probe in a second area of the display screen, the anatomy in the first image corresponding to the anatomy icon within the field of view.
2. The system of claim 1 , wherein the imaging system is an ultrasound imaging system, and the probe is an ultrasound probe.
3. The system of claim 1 , wherein the anatomy includes a heart.
4. The system of claim 1 , wherein the processor is configured to display the first image at a first orientation or a second orientation that is different than the first orientation without the probe being moved.
5. The system of claim 1 , wherein the processor is further configured to: display a second image of the anatomy in a third area of the display screen, the second image shows a different area of the anatomy than the first image.
6. The system of claim 5, wherein the second image is captured by the probe before or after the probe captures the first image; and wherein the processor is configured to display another field of view icon on the display corresponding to the second image.
PCT/IB2024/053600 2023-04-13 2024-04-12 System and method for illustrating a pose of an object WO2024214057A1 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US202363495951P 2023-04-13 2023-04-13
US63/495,951 2023-04-13
US18/630,474 US20240341860A1 (en) 2023-04-13 2024-04-09 System and method for illustrating a pose of an object
US18/630,474 2024-04-09

Publications (1)

Publication Number Publication Date
WO2024214057A1 true WO2024214057A1 (en) 2024-10-17

Family

ID=90825547

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2024/053600 WO2024214057A1 (en) 2023-04-13 2024-04-12 System and method for illustrating a pose of an object

Country Status (1)

Country Link
WO (1) WO2024214057A1 (en)

Citations (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5592939A (en) 1995-06-14 1997-01-14 Martinelli; Michael A. Method and system for navigating a catheter probe
US5913820A (en) 1992-08-14 1999-06-22 British Telecommunications Public Limited Company Position location system
US5983126A (en) 1995-11-22 1999-11-09 Medtronic, Inc. Catheter location system and method
US6379302B1 (en) 1999-10-28 2002-04-30 Surgical Navigation Technologies Inc. Navigation information overlay onto ultrasound imagery
US6474341B1 (en) 1999-10-28 2002-11-05 Surgical Navigation Technologies, Inc. Surgical communication and power system
US20030032878A1 (en) * 1996-06-28 2003-02-13 The Board Of Trustees Of The Leland Stanford Junior University Method and apparatus for volumetric image navigation
US6747539B1 (en) 1999-10-28 2004-06-08 Michael A. Martinelli Patient-shielding and coil system
US20040116803A1 (en) 2001-06-04 2004-06-17 Bradley Jascob Method and apparatus for electromagnetic navigation of a surgical probe near a metal object
US20040199072A1 (en) 2003-04-01 2004-10-07 Stacy Sprouse Integrated electromagnetic navigation and patient positioning device
US6940941B2 (en) 2002-02-15 2005-09-06 Breakaway Imaging, Llc Breakable gantry apparatus for multidimensional x-ray based imaging
US7001045B2 (en) 2002-06-11 2006-02-21 Breakaway Imaging, Llc Cantilevered gantry apparatus for x-ray imaging
US7085400B1 (en) 2000-06-14 2006-08-01 Surgical Navigation Technologies, Inc. System and method for image based sensor calibration
US7106825B2 (en) 2002-08-21 2006-09-12 Breakaway Imaging, Llc Apparatus and method for reconstruction of volumetric images in a divergent scanning computed tomography system
US7108421B2 (en) 2002-03-19 2006-09-19 Breakaway Imaging, Llc Systems and methods for imaging large field-of-view objects
US7188998B2 (en) 2002-03-13 2007-03-13 Breakaway Imaging, Llc Systems and methods for quasi-simultaneous multi-planar x-ray imaging
US7676268B2 (en) 2006-11-30 2010-03-09 Medtronic, Inc. Medical methods and systems incorporating wireless monitoring
US7697972B2 (en) 2002-11-19 2010-04-13 Medtronic Navigation, Inc. Navigation system for cardiac therapies
US7751865B2 (en) 2003-10-17 2010-07-06 Medtronic Navigation, Inc. Method and apparatus for surgical navigation
US20100228117A1 (en) 2009-03-09 2010-09-09 Medtronic Navigation, Inc System And Method For Image-Guided Navigation
US7797032B2 (en) 1999-10-28 2010-09-14 Medtronic Navigation, Inc. Method and system for navigating a catheter probe in the presence of field-influencing objects
US8060185B2 (en) 2002-11-19 2011-11-15 Medtronic Navigation, Inc. Navigation system for cardiac therapies
US8180428B2 (en) 2007-10-03 2012-05-15 Medtronic, Inc. Methods and systems for use in selecting cardiac pacing sites
US8238631B2 (en) 2009-05-13 2012-08-07 Medtronic Navigation, Inc. System and method for automatic registration between an image and a subject
US8494608B2 (en) 2008-04-18 2013-07-23 Medtronic, Inc. Method and apparatus for mapping a structure
US8532734B2 (en) 2008-04-18 2013-09-10 Regents Of The University Of Minnesota Method and apparatus for mapping a structure
US8644907B2 (en) 1999-10-28 2014-02-04 Medtronic Navigaton, Inc. Method and apparatus for surgical navigation
US8663120B2 (en) 2008-04-18 2014-03-04 Regents Of The University Of Minnesota Method and apparatus for mapping a structure
US8811662B2 (en) 2011-04-29 2014-08-19 Medtronic Navigation, Inc. Method and apparatus for calibrating and re-aligning an ultrasound image plane to a navigation tracker
US9138204B2 (en) 2011-04-29 2015-09-22 Medtronic Navigation, Inc. Method and apparatus for calibrating and re-aligning an ultrasound image plane to a navigation tracker
WO2017160889A1 (en) * 2016-03-14 2017-09-21 Mahfouz, Mohamed, R. Ultra-wideband positioning for wireless ultrasound tracking and communication
US20180078316A1 (en) 2016-09-22 2018-03-22 Medtronic Navigation, Inc. System for Guided Procedures
US20200159313A1 (en) * 2018-11-17 2020-05-21 Novarad Corporation Using Optical Codes with Augmented Reality Displays

Patent Citations (38)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5913820A (en) 1992-08-14 1999-06-22 British Telecommunications Public Limited Company Position location system
US5592939A (en) 1995-06-14 1997-01-14 Martinelli; Michael A. Method and system for navigating a catheter probe
US5983126A (en) 1995-11-22 1999-11-09 Medtronic, Inc. Catheter location system and method
US20030032878A1 (en) * 1996-06-28 2003-02-13 The Board Of Trustees Of The Leland Stanford Junior University Method and apparatus for volumetric image navigation
US6669635B2 (en) 1999-10-28 2003-12-30 Surgical Navigation Technologies, Inc. Navigation information overlay onto ultrasound imagery
US6474341B1 (en) 1999-10-28 2002-11-05 Surgical Navigation Technologies, Inc. Surgical communication and power system
US6747539B1 (en) 1999-10-28 2004-06-08 Michael A. Martinelli Patient-shielding and coil system
US8644907B2 (en) 1999-10-28 2014-02-04 Medtronic Navigaton, Inc. Method and apparatus for surgical navigation
US7797032B2 (en) 1999-10-28 2010-09-14 Medtronic Navigation, Inc. Method and system for navigating a catheter probe in the presence of field-influencing objects
US6968224B2 (en) 1999-10-28 2005-11-22 Surgical Navigation Technologies, Inc. Method of detecting organ matter shift in a patient
US6379302B1 (en) 1999-10-28 2002-04-30 Surgical Navigation Technologies Inc. Navigation information overlay onto ultrasound imagery
US8320653B2 (en) 2000-06-14 2012-11-27 Medtronic Navigation, Inc. System and method for image based sensor calibration
US7831082B2 (en) 2000-06-14 2010-11-09 Medtronic Navigation, Inc. System and method for image based sensor calibration
US7085400B1 (en) 2000-06-14 2006-08-01 Surgical Navigation Technologies, Inc. System and method for image based sensor calibration
US20040116803A1 (en) 2001-06-04 2004-06-17 Bradley Jascob Method and apparatus for electromagnetic navigation of a surgical probe near a metal object
US6940941B2 (en) 2002-02-15 2005-09-06 Breakaway Imaging, Llc Breakable gantry apparatus for multidimensional x-ray based imaging
US7188998B2 (en) 2002-03-13 2007-03-13 Breakaway Imaging, Llc Systems and methods for quasi-simultaneous multi-planar x-ray imaging
US7108421B2 (en) 2002-03-19 2006-09-19 Breakaway Imaging, Llc Systems and methods for imaging large field-of-view objects
US7001045B2 (en) 2002-06-11 2006-02-21 Breakaway Imaging, Llc Cantilevered gantry apparatus for x-ray imaging
US7106825B2 (en) 2002-08-21 2006-09-12 Breakaway Imaging, Llc Apparatus and method for reconstruction of volumetric images in a divergent scanning computed tomography system
US7697972B2 (en) 2002-11-19 2010-04-13 Medtronic Navigation, Inc. Navigation system for cardiac therapies
US8046052B2 (en) 2002-11-19 2011-10-25 Medtronic Navigation, Inc. Navigation system for cardiac therapies
US8060185B2 (en) 2002-11-19 2011-11-15 Medtronic Navigation, Inc. Navigation system for cardiac therapies
US20040199072A1 (en) 2003-04-01 2004-10-07 Stacy Sprouse Integrated electromagnetic navigation and patient positioning device
US7751865B2 (en) 2003-10-17 2010-07-06 Medtronic Navigation, Inc. Method and apparatus for surgical navigation
US7676268B2 (en) 2006-11-30 2010-03-09 Medtronic, Inc. Medical methods and systems incorporating wireless monitoring
US8180428B2 (en) 2007-10-03 2012-05-15 Medtronic, Inc. Methods and systems for use in selecting cardiac pacing sites
US8532734B2 (en) 2008-04-18 2013-09-10 Regents Of The University Of Minnesota Method and apparatus for mapping a structure
US8494608B2 (en) 2008-04-18 2013-07-23 Medtronic, Inc. Method and apparatus for mapping a structure
US8663120B2 (en) 2008-04-18 2014-03-04 Regents Of The University Of Minnesota Method and apparatus for mapping a structure
US20100228117A1 (en) 2009-03-09 2010-09-09 Medtronic Navigation, Inc System And Method For Image-Guided Navigation
US9737235B2 (en) 2009-03-09 2017-08-22 Medtronic Navigation, Inc. System and method for image-guided navigation
US8238631B2 (en) 2009-05-13 2012-08-07 Medtronic Navigation, Inc. System and method for automatic registration between an image and a subject
US8811662B2 (en) 2011-04-29 2014-08-19 Medtronic Navigation, Inc. Method and apparatus for calibrating and re-aligning an ultrasound image plane to a navigation tracker
US9138204B2 (en) 2011-04-29 2015-09-22 Medtronic Navigation, Inc. Method and apparatus for calibrating and re-aligning an ultrasound image plane to a navigation tracker
WO2017160889A1 (en) * 2016-03-14 2017-09-21 Mahfouz, Mohamed, R. Ultra-wideband positioning for wireless ultrasound tracking and communication
US20180078316A1 (en) 2016-09-22 2018-03-22 Medtronic Navigation, Inc. System for Guided Procedures
US20200159313A1 (en) * 2018-11-17 2020-05-21 Novarad Corporation Using Optical Codes with Augmented Reality Displays

Similar Documents

Publication Publication Date Title
US10838577B2 (en) Method and apparatus for image-based navigation
US9597154B2 (en) Method and apparatus for optimizing a computer assisted surgical procedure
US8364242B2 (en) System and method of combining ultrasound image acquisition with fluoroscopic image acquisition
US8428690B2 (en) Intracardiac echocardiography image reconstruction in combination with position tracking system
EP2331001B1 (en) System for tracking a patient
EP1583469B1 (en) Method and system for registering a medical situation associated with a first coordinate system, in a second coordinate system using an mps system
US8010177B2 (en) Intraoperative image registration
EP1912565B1 (en) Catheter navigation system
US20240341860A1 (en) System and method for illustrating a pose of an object
US20080009731A1 (en) Radiotherapy device
JP2002153443A (en) Method and apparatus for mapping cardiac chamber
WO2011053432A1 (en) System and method to evaluate electrode position and spacing
WO2008130354A1 (en) Intraoperative image registration
EP3675039B1 (en) Integration of medical imaging and location tracking
US20140316256A1 (en) Display Of An Acquired Cine Loop For Procedure Navigation
EP2432388B1 (en) System for cardiac lead placement
WO2024214057A1 (en) System and method for illustrating a pose of an object
WO2024089504A1 (en) System operable to determine a pose of an instrument
WO2024089502A1 (en) System and method for illustrating a pose of an object
WO2024089503A1 (en) System and method for illustrating a pose of an object
WO2024215790A1 (en) System and method for positioning a member relative to a subject
WO2024215791A1 (en) System and method for positioning a member relative to a subject
WO2024157104A1 (en) Tissue property localization for therapy delivery

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 24720604

Country of ref document: EP

Kind code of ref document: A1