CA2700577A1 - Mri surgical systems for real-time visualizations using mri image data and predefined data of surgical tools - Google Patents
Mri surgical systems for real-time visualizations using mri image data and predefined data of surgical tools Download PDFInfo
- Publication number
- CA2700577A1 CA2700577A1 CA2700577A CA2700577A CA2700577A1 CA 2700577 A1 CA2700577 A1 CA 2700577A1 CA 2700577 A CA2700577 A CA 2700577A CA 2700577 A CA2700577 A CA 2700577A CA 2700577 A1 CA2700577 A1 CA 2700577A1
- Authority
- CA
- Canada
- Prior art keywords
- circuit
- patient
- mri
- trajectory
- tool
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01R—MEASURING ELECTRIC VARIABLES; MEASURING MAGNETIC VARIABLES
- G01R33/00—Arrangements or instruments for measuring magnetic variables
- G01R33/20—Arrangements or instruments for measuring magnetic variables involving magnetic resonance
- G01R33/28—Details of apparatus provided for in groups G01R33/44 - G01R33/64
- G01R33/283—Intercom or optical viewing arrangements, structurally associated with NMR apparatus
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/70—Manipulators specially adapted for use in surgery
- A61B34/71—Manipulators operated by drive cable mechanisms
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/70—Manipulators specially adapted for use in surgery
- A61B34/74—Manipulators with manual electric input means
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/05—Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves
- A61B5/055—Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves involving electronic [EMR] or nuclear [NMR] magnetic resonance, e.g. magnetic resonance imaging
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/10—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges for stereotaxic surgery, e.g. frame-based stereotaxis
- A61B90/11—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges for stereotaxic surgery, e.g. frame-based stereotaxis with guides for needles or instruments, e.g. arcuate slides or ball joints
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01R—MEASURING ELECTRIC VARIABLES; MEASURING MAGNETIC VARIABLES
- G01R33/00—Arrangements or instruments for measuring magnetic variables
- G01R33/20—Arrangements or instruments for measuring magnetic variables involving magnetic resonance
- G01R33/28—Details of apparatus provided for in groups G01R33/44 - G01R33/64
- G01R33/285—Invasive instruments, e.g. catheters or biopsy needles, specially adapted for tracking, guiding or visualization by NMR
- G01R33/286—Invasive instruments, e.g. catheters or biopsy needles, specially adapted for tracking, guiding or visualization by NMR involving passive visualization of interventional instruments, i.e. making the instrument visible as part of the normal MR process
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01R—MEASURING ELECTRIC VARIABLES; MEASURING MAGNETIC VARIABLES
- G01R33/00—Arrangements or instruments for measuring magnetic variables
- G01R33/20—Arrangements or instruments for measuring magnetic variables involving magnetic resonance
- G01R33/28—Details of apparatus provided for in groups G01R33/44 - G01R33/64
- G01R33/285—Invasive instruments, e.g. catheters or biopsy needles, specially adapted for tracking, guiding or visualization by NMR
- G01R33/287—Invasive instruments, e.g. catheters or biopsy needles, specially adapted for tracking, guiding or visualization by NMR involving active visualization of interventional instruments, e.g. using active tracking RF coils or coils for intentionally creating magnetic field inhomogeneities
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01R—MEASURING ELECTRIC VARIABLES; MEASURING MAGNETIC VARIABLES
- G01R33/00—Arrangements or instruments for measuring magnetic variables
- G01R33/20—Arrangements or instruments for measuring magnetic variables involving magnetic resonance
- G01R33/28—Details of apparatus provided for in groups G01R33/44 - G01R33/64
- G01R33/32—Excitation or detection systems, e.g. using radio frequency signals
- G01R33/34—Constructional details, e.g. resonators, specially adapted to MR
- G01R33/34046—Volume type coils, e.g. bird-cage coils; Quadrature bird-cage coils; Circularly polarised coils
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01R—MEASURING ELECTRIC VARIABLES; MEASURING MAGNETIC VARIABLES
- G01R33/00—Arrangements or instruments for measuring magnetic variables
- G01R33/20—Arrangements or instruments for measuring magnetic variables involving magnetic resonance
- G01R33/28—Details of apparatus provided for in groups G01R33/44 - G01R33/64
- G01R33/32—Excitation or detection systems, e.g. using radio frequency signals
- G01R33/34—Constructional details, e.g. resonators, specially adapted to MR
- G01R33/341—Constructional details, e.g. resonators, specially adapted to MR comprising surface coils
- G01R33/3415—Constructional details, e.g. resonators, specially adapted to MR comprising surface coils comprising arrays of sub-coils, i.e. phased-array coils with flexible receiver channels
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B2017/00017—Electrical control of surgical instruments
- A61B2017/00212—Electrical control of surgical instruments using remote controls
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B2017/00831—Material properties
- A61B2017/00902—Material properties transparent or translucent
- A61B2017/00911—Material properties transparent or translucent for fields applied by a magnetic resonance imaging system
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B17/34—Trocars; Puncturing needles
- A61B17/3403—Needle locating or guiding means
- A61B2017/3405—Needle locating or guiding means using mechanical guide means
- A61B2017/3407—Needle locating or guiding means using mechanical guide means including a base for support on the body
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B17/34—Trocars; Puncturing needles
- A61B17/3403—Needle locating or guiding means
- A61B2017/3405—Needle locating or guiding means using mechanical guide means
- A61B2017/3409—Needle locating or guiding means using mechanical guide means including needle or instrument drives
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
- A61B2034/101—Computer-aided simulation of surgical operations
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
- A61B2034/107—Visualisation of planned trajectories or target regions
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/25—User interfaces for surgical systems
- A61B2034/252—User interfaces for surgical systems indicating steps of a surgical procedure
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/25—User interfaces for surgical systems
- A61B2034/254—User interfaces for surgical systems being adapted depending on the stage of the surgical procedure
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/25—User interfaces for surgical systems
- A61B2034/256—User interfaces for surgical systems having a database of accessory information, e.g. including context sensitive help or scientific articles
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/10—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges for stereotaxic surgery, e.g. frame-based stereotaxis
- A61B2090/103—Cranial plugs for access to brain
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/30—Devices for illuminating a surgical field, the devices having an interrelation with other surgical devices or with a surgical procedure
- A61B2090/306—Devices for illuminating a surgical field, the devices having an interrelation with other surgical devices or with a surgical procedure using optical fibres
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B90/361—Image-producing devices, e.g. surgical cameras
- A61B2090/3614—Image-producing devices, e.g. surgical cameras using optical fibre
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B90/37—Surgical systems with images on a monitor during operation
- A61B2090/373—Surgical systems with images on a monitor during operation using light, e.g. by using optical scanners
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B90/37—Surgical systems with images on a monitor during operation
- A61B2090/374—NMR or MRI
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/39—Markers, e.g. radio-opaque or breast lesions markers
- A61B2090/3933—Liquid markers
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/39—Markers, e.g. radio-opaque or breast lesions markers
- A61B2090/3937—Visible markers
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/39—Markers, e.g. radio-opaque or breast lesions markers
- A61B2090/3937—Visible markers
- A61B2090/395—Visible markers with marking agent for marking skin or other tissue
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/39—Markers, e.g. radio-opaque or breast lesions markers
- A61B2090/3954—Markers, e.g. radio-opaque or breast lesions markers magnetic, e.g. NMR or MRI
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/39—Markers, e.g. radio-opaque or breast lesions markers
- A61B2090/3983—Reference marker arrangements for use with image guided surgery
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/25—User interfaces for surgical systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/40—Detecting, measuring or recording for evaluating the nervous system
- A61B5/4058—Detecting, measuring or recording for evaluating the nervous system for evaluating the central nervous system
- A61B5/4064—Evaluating the brain
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/10—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges for stereotaxic surgery, e.g. frame-based stereotaxis
- A61B90/14—Fixators for body parts, e.g. skull clamps; Constructional details of fixators, e.g. pins
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B90/361—Image-producing devices, e.g. surgical cameras
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B90/37—Surgical systems with images on a monitor during operation
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61N—ELECTROTHERAPY; MAGNETOTHERAPY; RADIATION THERAPY; ULTRASOUND THERAPY
- A61N1/00—Electrotherapy; Circuits therefor
- A61N1/18—Applying electric currents by contact electrodes
- A61N1/32—Applying electric currents by contact electrodes alternating or intermittent currents
- A61N1/36—Applying electric currents by contact electrodes alternating or intermittent currents for stimulation
- A61N1/372—Arrangements in connection with the implantation of stimulators
Abstract
MRI-Surgical systems include: (a) at least one MRI-compatible surgical tool;
(b) a circuit adapted to communicate with an MRI scanner; and (c) at least one display in communication with the circuit. The circuit electronically recognizes predefined physical characteristics of the at least one tool to automatically segment MR
image data provided by the MRI scanner whereby the at least one tool constitutes a point of interface with the system. The circuit is configured to provide a User Interface that defines workflow progression for an MRI-guided surgical procedure and allows a user to select steps in the workflow, and wherein the circuit is configured to generate multi-dimensional visualizations using the predefined data of the at least one tool and data from MRI images of the patient in substantially real time during the surgical procedure.
(b) a circuit adapted to communicate with an MRI scanner; and (c) at least one display in communication with the circuit. The circuit electronically recognizes predefined physical characteristics of the at least one tool to automatically segment MR
image data provided by the MRI scanner whereby the at least one tool constitutes a point of interface with the system. The circuit is configured to provide a User Interface that defines workflow progression for an MRI-guided surgical procedure and allows a user to select steps in the workflow, and wherein the circuit is configured to generate multi-dimensional visualizations using the predefined data of the at least one tool and data from MRI images of the patient in substantially real time during the surgical procedure.
Description
MRI SURGICAL SYSTEMS FOR REAL-TIME VISUALIZATIONS USING
MRI IMAGE DATA AND PREDEFINED DATA OF SURGICAL TOOLS
RELATED APPLICATIONS
[0001] This application claims priority to U.S. Application Serial No.
12/134,412, filed June 6, 2008, and also claims priority to U.S. Provisional Application Serial No. 60/974,821, filed September 24, 2007, the contents of which are hereby incorporated by reference as if recited in full herein.
FIELD OF THE INVENTION
MRI IMAGE DATA AND PREDEFINED DATA OF SURGICAL TOOLS
RELATED APPLICATIONS
[0001] This application claims priority to U.S. Application Serial No.
12/134,412, filed June 6, 2008, and also claims priority to U.S. Provisional Application Serial No. 60/974,821, filed September 24, 2007, the contents of which are hereby incorporated by reference as if recited in full herein.
FIELD OF THE INVENTION
[0002] The present invention relates to MRI-guided diagnostic or interventional systems that may be particularly suitable for placement/localization of interventional medical devices and/or therapies in the body. Embodiments of the present invention may be particularly suitable for placing neuromodulation leads, such as Deep Brain Stimulation ("DBS") leads, placing implantable parasympathetic or sympathetic nerve chain leads and/or CNS stimulation leads and/or for delivering therapies to target internal locations in the body including atrial fibrillation (AFIB) therapies.
BACKGROUND OF THE INVENTION
BACKGROUND OF THE INVENTION
[0003] Deep Brain Stimulation (DBS) is becoming an acceptable therapeutic modality in neurosurgical treatment of patients suffering from chronic pain, Parkinson's disease or seizure, and other medical conditions. Other electro-stimulation therapies have also been carried out or proposed using internal stimulation of the sympathetic nerve chain and/or spinal cord, etc.
[0004] One example of a prior art DBS system is the Activa system from Medtronic, Inc. The Activa system includes an implantable pulse generator stimulator that is positioned in the chest cavity of the patient and a lead with axially spaced apart electrodes that is implanted with the electrodes disposed in neural tissue.
The lead is tunneled subsurface from the brain to the chest cavity connecting the electrodes with the pulse generator. These leads can have multiple exposed electrodes at the distal end that are connected to conductors which run along the length of the
The lead is tunneled subsurface from the brain to the chest cavity connecting the electrodes with the pulse generator. These leads can have multiple exposed electrodes at the distal end that are connected to conductors which run along the length of the
5 PCT/US2008/011050 lead and connect to the pulse generator placed in the chest cavity.
[0005] It is believed that the clinical outcome of certain medical procedures, particularly those using DBS, may depend on the precise location of the electrodes that are in contact with the tissue of interest. For example, to treat Parkinson's tremor, DBS stimulation leads are conventionally implanted during a stereotactic surgery, based on pre-operative MRI and CT images. These procedures can be long in duration and may have reduced efficacy as it has been reported that, in about 30% of the patients implanted with these devices, the clinical efficacy of the device/procedure is less than optimum.
SUMMARY OF EMBODIMENTS OF THE INVENTION
[00061 Some embodiments of the present invention are directed to MRI-guided systems that can generate substantially real time patient-specific visualizations of the patient and one or more surgical tools in logical space and provide feedback to a clinician to improve the speed and/or reliability of an intrabody procedure.
[0007] The visualizations can be based (in-part) on predefined data of the tool(s) which define a point of interface for the system (e.g., software) based on predefined characteristics of the tool(s), e.g., dimensions, shape or configuration and/or known rotational, translational and/or other functional and/or dynamic behavior of one or more surgical tools. The visualizations can include patient function data (e.g., fMRI data, electrical activity, active regions of a brain during a defined stimulation, fiber tracks, and the like).
[00081 The system can be configured to interrogate and segment image data to locate fiducial markers and generate successive visualizations of the patient's anatomical structure and tool(s) using MRI image data and apriori data of the tool(s) to provide (substantially real-time) visualizations of the patient.
[0009] Some embodiments are directed to MRI-guided surgical systems.
The systems include: (a) at least one MRI-compatible surgical tool; (b) a circuit adapted to communicate with an MRI scanner; and (c) at least one display in communication with the circuit. The circuit electronically recognizes predefined physical characteristics of the at least one tool to automatically segment MR
image data provided by the MRI scanner whereby the at least one tool constitutes a point of interface with the system. The circuit is configured to provide a User Interface that defines workflow progression for an MRI-guided surgical procedure and allows a user to select steps in the workflow, and wherein the circuit is configured to generate multi-dimensional visualizations using the predefined data of the at least one tool and data from MRI images of the patient in substantially real time during the surgical procedure.
[0010) Other embodiments are directed to methods for performing an MRI-guided surgical procedure. The methods include: (a) defining dimensional and/or functional data of at least one MRI compatible surgical tool; (b) obtaining MRI
image data of the patient; (c) electronically segmenting the MRI image data to identify known fiducial markers on the at least one tool based on the defining step; (d) generating visualizations of the at least one tool registered to patient anatomical structure; (e) electronically generating directions on adjustments for a pitch, roll or X-Y actuator to adjust a trajectory of a trajectory guide; and (f) guiding the tool to a location in the patient using patient MRI image data, the directions for adjustment and the visualizations thereby facilitating an MRI-guided surgical procedure.
[0011] Still other embodiments are directed to computer program products for facilitating an MRI-guided surgical procedure. The computer program product includes a computer readable storage medium having computer readable program code embodied in the medium. The computer-readable program code includes: (a) computer readable program code that comprises predefined physical data of a plurality of different surgical tools; (b) computer readable program code that communicates with an MRI scanner to obtain MRI image data of a patient; and (c) computer readable program code that generates visualizations of the patient using data from the tools and the image data of the patient in substantially real-time.
[0012] Yet other embodiments are directed to MRI-guided interventional deep brain systems. The systems include: (a) an MRI Scanner; (b) a clinician workstation with a circuit and a display, the workstation in communication with the MRI Scanner; (c) at least one flexible patch with a grid thereon configured to releasably attach to a patient's skull; and (d) at least one trajectory guide attachable to a skull of a patient. The guide has a base with an aperture configured to reside over a burr hole formed in a patient's skull. The base aperture provides a mechanical center of rotation for a pivot axis associated with the trajectory guide, the base having a plurality of fiducial markers spaced apart about the base aperture. The circuit comprises physical data regarding the patch and is configured to interrogate patient imaging data provided by the MRI Scanner and segment the image data to define a burr hole location that intersects the patch with a desired intrabrain trajectory. The circuit comprises tool-specific data of the trajectory guide and is configured to interrogate patient imaging data provided by the MRI Scanner and interactively generate''visualizations of the patient's brain and the trajectory guide to the display.
[0013] In some embodiments, the circuit is configured to provide a default trajectory for the trajectory guide on the display that extends through a center location of the grid patch.
[0014] Embodiments of the invention can provide output to a user such as one or more of: (a) electronic generated warnings to alert an improper planned trajectory for a trajectory guide; (b) warnings regarding a physical interference with a planned projected trajectory associated with the MRI bore size and (isocenter) position (and optionally, patient head size and angle(s) or configuration of a surgical tool); (c) electronic instructions on what grid entry location to use to obtain a desired trajectory or entry point into the patient brain; (d) calculate and provide suggested physical adjustments to actuators to obtain a desired trajectory orientation and generate instructions on what adjustments to make to X, Y, pitch and roll adjustment mechanisms (e.g., rotate X button or dial left or right, potentially with a number of rotations or increments and the like) associated with the trajectory guide to obtain the desired trajectory; and (e) generate electronic data of electrode offset values for stimulation leads in the brain to define where the electrodes are anatomically positioned whereby pulse generator programming may be accelerated over conventional techniques.
[0015] Some embodiments of the present invention can provide visualizations to allow more precise control, delivery, and/or feedback of a therapy so that the therapy or a tool associated therewith can be more precisely placed, delivered, confirmed and visualized.
[0016] These and other embodiments will be described further below.
BRIEF DESCRIPTION OF THE DRAWINGS
[0017] Figure 1 is a schematic illustration of a MRI-guided surgical system according to some embodiments of the present invention.
[0018] Figure 2 is a schematic illustration of an MRI-guided surgical system with MRI compatible cameras according to embodiments of the present invention.
[0019] Figure 3 is a schematic illustration of an MRI-guided surgical system according to some embodiments of the present invention.
[0020] Figure 4 is a schematic illustration of an exemplary screen shot of a user interface according to some embodiments of the present invention.
[0021] Figure 5 is a schematic of exemplary disposable hardware that can be used to carryout embodiments of the present invention.
[0022] Figure 6A is a schematic of an exemplary trajectory guide in position on a patient according to some embodiments of the present invention.
[0023] Figure 6B is a side view of a depth stop with a cooperating elongate member according to some embodiments of the present invention.
[0024] Figure 6C is a side view of a depth stop cooperating with an elongate member and peel-away sheath according to some embodiments of the present invention.
[0025] Figure 6D is a side perspective view of the depth stop and sheath cooperating with the trajectory guide according to some embodiments of the present invention.
[0026] Figure 7 is a side perspective view of a trajectory guide and optional camera device according to some embodiments of the present invention.
[0027] Figure 8 is a sectional view of the trajectory guide with a targeting canula according to some embodiments of the present invention.
[0028] Figure 9 is a top view of a base of a trajectory guide with fiducials according to some embodiments of the present invention.
[0029] Figure 10 is a side perspective view of the base shown in Figure 9.
[0030] Figure 11 is a schematic illustration of a marking grid patch and associated screen display of coordinates of a surgical entry site according to embodiinents of the present invention.
[0031] Figures 12A-12D are schematic illustrations of grid segmentation and grid deformation that can be used to define an entry site location according to embodiments of the present invention.
[0032] Figures 13A-13C are schematic illustrations of a base or frame marker segmentation that can be used to define position and orientation of a base or frame of a trajectory guide according to embodiments of the present invention.
[0033] Figures 14A and 14B are illustrations of a User Interface (UI) tool bar with exemplary workflow groups according to embodiments of the present invention. Figure 14B illustrates UI selectable steps for a selected workflow group according to some embodiments of the present invention.
[0034] Figure 15 is a screen shot of an exemplary (e.g., DBS) workstation Start window for a workstation display according to embodiments of the present invention.
[0035] Figures 16-19 and 22-38 are screen shots of exemplary displays of different workflow groups and/or steps associated with a User Interface provided to a user to facilitate an MRI-guided procedure.
[00361 Figures 20 and 21 are examples of operational warnings provided to a workstation/display that can be automatically generated by the system according to some embodiments of the present invention.
[0037] Figure 39 is a data processing system according to some embodiments of the present invention.
DETAILED DESCRIPTION OF EMBODIMENTS OF THE INVENTION
[0038] The present invention will now be described more fully hereinafter with reference to the accompanying drawings, in which embodiments of the invention are shown. This invention may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein;
rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the invention to those skilled in the art. Like numbers refer to like elements throughout. It will be appreciated that although discussed with respect to a certain embodiment, features or operation of one embodiment can apply to others.
[0039] In the drawings, the thickness of lines, layers, features, components and/or regions may be exaggerated for clarity and broken lines (such as those shown in circuit of flow diagrams) illustrate optional features or operations, unless specified otherwise. In addition, the sequence of operations (or steps) is not limited to the order presented in the claims unless specifically indicated otherwise.
[0040] The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used herein, the singular forms "a", "an" and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms "comprises" and/or "comprising," when used in this
[0005] It is believed that the clinical outcome of certain medical procedures, particularly those using DBS, may depend on the precise location of the electrodes that are in contact with the tissue of interest. For example, to treat Parkinson's tremor, DBS stimulation leads are conventionally implanted during a stereotactic surgery, based on pre-operative MRI and CT images. These procedures can be long in duration and may have reduced efficacy as it has been reported that, in about 30% of the patients implanted with these devices, the clinical efficacy of the device/procedure is less than optimum.
SUMMARY OF EMBODIMENTS OF THE INVENTION
[00061 Some embodiments of the present invention are directed to MRI-guided systems that can generate substantially real time patient-specific visualizations of the patient and one or more surgical tools in logical space and provide feedback to a clinician to improve the speed and/or reliability of an intrabody procedure.
[0007] The visualizations can be based (in-part) on predefined data of the tool(s) which define a point of interface for the system (e.g., software) based on predefined characteristics of the tool(s), e.g., dimensions, shape or configuration and/or known rotational, translational and/or other functional and/or dynamic behavior of one or more surgical tools. The visualizations can include patient function data (e.g., fMRI data, electrical activity, active regions of a brain during a defined stimulation, fiber tracks, and the like).
[00081 The system can be configured to interrogate and segment image data to locate fiducial markers and generate successive visualizations of the patient's anatomical structure and tool(s) using MRI image data and apriori data of the tool(s) to provide (substantially real-time) visualizations of the patient.
[0009] Some embodiments are directed to MRI-guided surgical systems.
The systems include: (a) at least one MRI-compatible surgical tool; (b) a circuit adapted to communicate with an MRI scanner; and (c) at least one display in communication with the circuit. The circuit electronically recognizes predefined physical characteristics of the at least one tool to automatically segment MR
image data provided by the MRI scanner whereby the at least one tool constitutes a point of interface with the system. The circuit is configured to provide a User Interface that defines workflow progression for an MRI-guided surgical procedure and allows a user to select steps in the workflow, and wherein the circuit is configured to generate multi-dimensional visualizations using the predefined data of the at least one tool and data from MRI images of the patient in substantially real time during the surgical procedure.
[0010) Other embodiments are directed to methods for performing an MRI-guided surgical procedure. The methods include: (a) defining dimensional and/or functional data of at least one MRI compatible surgical tool; (b) obtaining MRI
image data of the patient; (c) electronically segmenting the MRI image data to identify known fiducial markers on the at least one tool based on the defining step; (d) generating visualizations of the at least one tool registered to patient anatomical structure; (e) electronically generating directions on adjustments for a pitch, roll or X-Y actuator to adjust a trajectory of a trajectory guide; and (f) guiding the tool to a location in the patient using patient MRI image data, the directions for adjustment and the visualizations thereby facilitating an MRI-guided surgical procedure.
[0011] Still other embodiments are directed to computer program products for facilitating an MRI-guided surgical procedure. The computer program product includes a computer readable storage medium having computer readable program code embodied in the medium. The computer-readable program code includes: (a) computer readable program code that comprises predefined physical data of a plurality of different surgical tools; (b) computer readable program code that communicates with an MRI scanner to obtain MRI image data of a patient; and (c) computer readable program code that generates visualizations of the patient using data from the tools and the image data of the patient in substantially real-time.
[0012] Yet other embodiments are directed to MRI-guided interventional deep brain systems. The systems include: (a) an MRI Scanner; (b) a clinician workstation with a circuit and a display, the workstation in communication with the MRI Scanner; (c) at least one flexible patch with a grid thereon configured to releasably attach to a patient's skull; and (d) at least one trajectory guide attachable to a skull of a patient. The guide has a base with an aperture configured to reside over a burr hole formed in a patient's skull. The base aperture provides a mechanical center of rotation for a pivot axis associated with the trajectory guide, the base having a plurality of fiducial markers spaced apart about the base aperture. The circuit comprises physical data regarding the patch and is configured to interrogate patient imaging data provided by the MRI Scanner and segment the image data to define a burr hole location that intersects the patch with a desired intrabrain trajectory. The circuit comprises tool-specific data of the trajectory guide and is configured to interrogate patient imaging data provided by the MRI Scanner and interactively generate''visualizations of the patient's brain and the trajectory guide to the display.
[0013] In some embodiments, the circuit is configured to provide a default trajectory for the trajectory guide on the display that extends through a center location of the grid patch.
[0014] Embodiments of the invention can provide output to a user such as one or more of: (a) electronic generated warnings to alert an improper planned trajectory for a trajectory guide; (b) warnings regarding a physical interference with a planned projected trajectory associated with the MRI bore size and (isocenter) position (and optionally, patient head size and angle(s) or configuration of a surgical tool); (c) electronic instructions on what grid entry location to use to obtain a desired trajectory or entry point into the patient brain; (d) calculate and provide suggested physical adjustments to actuators to obtain a desired trajectory orientation and generate instructions on what adjustments to make to X, Y, pitch and roll adjustment mechanisms (e.g., rotate X button or dial left or right, potentially with a number of rotations or increments and the like) associated with the trajectory guide to obtain the desired trajectory; and (e) generate electronic data of electrode offset values for stimulation leads in the brain to define where the electrodes are anatomically positioned whereby pulse generator programming may be accelerated over conventional techniques.
[0015] Some embodiments of the present invention can provide visualizations to allow more precise control, delivery, and/or feedback of a therapy so that the therapy or a tool associated therewith can be more precisely placed, delivered, confirmed and visualized.
[0016] These and other embodiments will be described further below.
BRIEF DESCRIPTION OF THE DRAWINGS
[0017] Figure 1 is a schematic illustration of a MRI-guided surgical system according to some embodiments of the present invention.
[0018] Figure 2 is a schematic illustration of an MRI-guided surgical system with MRI compatible cameras according to embodiments of the present invention.
[0019] Figure 3 is a schematic illustration of an MRI-guided surgical system according to some embodiments of the present invention.
[0020] Figure 4 is a schematic illustration of an exemplary screen shot of a user interface according to some embodiments of the present invention.
[0021] Figure 5 is a schematic of exemplary disposable hardware that can be used to carryout embodiments of the present invention.
[0022] Figure 6A is a schematic of an exemplary trajectory guide in position on a patient according to some embodiments of the present invention.
[0023] Figure 6B is a side view of a depth stop with a cooperating elongate member according to some embodiments of the present invention.
[0024] Figure 6C is a side view of a depth stop cooperating with an elongate member and peel-away sheath according to some embodiments of the present invention.
[0025] Figure 6D is a side perspective view of the depth stop and sheath cooperating with the trajectory guide according to some embodiments of the present invention.
[0026] Figure 7 is a side perspective view of a trajectory guide and optional camera device according to some embodiments of the present invention.
[0027] Figure 8 is a sectional view of the trajectory guide with a targeting canula according to some embodiments of the present invention.
[0028] Figure 9 is a top view of a base of a trajectory guide with fiducials according to some embodiments of the present invention.
[0029] Figure 10 is a side perspective view of the base shown in Figure 9.
[0030] Figure 11 is a schematic illustration of a marking grid patch and associated screen display of coordinates of a surgical entry site according to embodiinents of the present invention.
[0031] Figures 12A-12D are schematic illustrations of grid segmentation and grid deformation that can be used to define an entry site location according to embodiments of the present invention.
[0032] Figures 13A-13C are schematic illustrations of a base or frame marker segmentation that can be used to define position and orientation of a base or frame of a trajectory guide according to embodiments of the present invention.
[0033] Figures 14A and 14B are illustrations of a User Interface (UI) tool bar with exemplary workflow groups according to embodiments of the present invention. Figure 14B illustrates UI selectable steps for a selected workflow group according to some embodiments of the present invention.
[0034] Figure 15 is a screen shot of an exemplary (e.g., DBS) workstation Start window for a workstation display according to embodiments of the present invention.
[0035] Figures 16-19 and 22-38 are screen shots of exemplary displays of different workflow groups and/or steps associated with a User Interface provided to a user to facilitate an MRI-guided procedure.
[00361 Figures 20 and 21 are examples of operational warnings provided to a workstation/display that can be automatically generated by the system according to some embodiments of the present invention.
[0037] Figure 39 is a data processing system according to some embodiments of the present invention.
DETAILED DESCRIPTION OF EMBODIMENTS OF THE INVENTION
[0038] The present invention will now be described more fully hereinafter with reference to the accompanying drawings, in which embodiments of the invention are shown. This invention may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein;
rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the invention to those skilled in the art. Like numbers refer to like elements throughout. It will be appreciated that although discussed with respect to a certain embodiment, features or operation of one embodiment can apply to others.
[0039] In the drawings, the thickness of lines, layers, features, components and/or regions may be exaggerated for clarity and broken lines (such as those shown in circuit of flow diagrams) illustrate optional features or operations, unless specified otherwise. In addition, the sequence of operations (or steps) is not limited to the order presented in the claims unless specifically indicated otherwise.
[0040] The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used herein, the singular forms "a", "an" and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms "comprises" and/or "comprising," when used in this
6 specification, specify the presence of stated features, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, steps, operations, elements, components, and/or groups thereof. As used herein, the term "and/or" includes any and all combinations of one or more of the associated listed items.
[0041] Unless otherwise defined, all terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. It will be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the specification and relevant art and should not be interpreted in an idealized or overly formal sense unless expressly so defined herein. Well-known functions or constructions may not be described in detail for brevity and/or clarity.
[0042] It will be understood that when a feature, such as a layer, region or substrate, is referred to as being "on" another feature or element, it can be directly on the other element or intervening elements may also be present. In contrast, when an element is referred to as being "directly on" another feature or element, there are no intervening elements present. It will also be understood that, when a feature or element is referred to as being "connected" or "coupled" to another feature or element, it can be directly connected to the other element or intervening elements may be present. In contrast, when a feature or element is referred to as being "directly connected" or "directly coupled" to another element, there are no intervening elements present. Although described or shown with respect to one embodiment, the features so described or shown can apply to other embodiments.
[0043] The term "electroanatomical visualization" or refers to a visualization or map of the anatomical structure, e.g., brain or heart, typically a volumetric, 3-D map or 4-D map, that illustrates or shows electrical activity of tissue correlated to anatomical and/or coordinate spatial position. The visualization can be in color and color-coded to provide an easy to understand map or image with different measures or gradients of activity in different colors and/or intensities. The term "color-coded" means that certain features electrical activity or other output are shown with defined colors of different color and/or intensity to visually accentuate different tissue, different and similar electrical activity or potential in tissue and/or to show abnormalities or lesions in tissue versus normal or non-lesion tissue. In some
[0041] Unless otherwise defined, all terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. It will be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the specification and relevant art and should not be interpreted in an idealized or overly formal sense unless expressly so defined herein. Well-known functions or constructions may not be described in detail for brevity and/or clarity.
[0042] It will be understood that when a feature, such as a layer, region or substrate, is referred to as being "on" another feature or element, it can be directly on the other element or intervening elements may also be present. In contrast, when an element is referred to as being "directly on" another feature or element, there are no intervening elements present. It will also be understood that, when a feature or element is referred to as being "connected" or "coupled" to another feature or element, it can be directly connected to the other element or intervening elements may be present. In contrast, when a feature or element is referred to as being "directly connected" or "directly coupled" to another element, there are no intervening elements present. Although described or shown with respect to one embodiment, the features so described or shown can apply to other embodiments.
[0043] The term "electroanatomical visualization" or refers to a visualization or map of the anatomical structure, e.g., brain or heart, typically a volumetric, 3-D map or 4-D map, that illustrates or shows electrical activity of tissue correlated to anatomical and/or coordinate spatial position. The visualization can be in color and color-coded to provide an easy to understand map or image with different measures or gradients of activity in different colors and/or intensities. The term "color-coded" means that certain features electrical activity or other output are shown with defined colors of different color and/or intensity to visually accentuate different tissue, different and similar electrical activity or potential in tissue and/or to show abnormalities or lesions in tissue versus normal or non-lesion tissue. In some
7 embodiments, the systems can be configured to allow a clinician to increase or decrease the intensity or change a color of certain tissue types or electrical outputs, e.g., in high-contrast color and/or intensity, darker opacity or the like.
(0044] The actual visualization can be shown on a screen or display so that the map and/or anatomical or tool structure is in a flat 2-D view and/or in 2-D what appears to be 3-D volumetric images with data representing features or electrical output with different visual characteristics such as with differing intensity, opacity, color, texture and the like. A 4-D map illustrates time-dependent activity, such as electrical activity or blood flow movement.
[0045] The systems are configured to operate based on known physical characteristics of one or more surgical tools such that the hardware is a point of interface for the circuit or software. The systems can communicate with databases that define dimensions, configurations or shapes and spacing of components on the tool(s). The defined physical data can be obtained from a CAD model of a tool.
The physical characteristics can include dimensions or other physical features or attributes and may also include relative changes in position of certain components or features upon a change in position of a tool or portion thereof. The defined physical characteristics can be electronically (programmatically) accessible by the system or known a priori and electronically stored locally or remotely and used to automatically calculate certain information and/or to segment image data. That is, the tool data from the model can be used to segment image data and/or correlate a position and orientation of a tool and/or provide trajectory adjustment guidelines or error estimates, warnings of improper trajectories and the like. For example, a grid for marking a burr hole location and/or a trajectory guide that adjusts an intrabrain path for placing a diagnostic or therapy device and such can be input, transposed, and/or overlayed in a visualization of the tool and patient structure or otherwise used, such as, for example, to project the information onto a patient's anatomical structure or determine certain operational parameters including which image volume to obtain high resolution MRI
image data that include select portions of the targeting canula. At least some of the resulting visualizations are not merely an MRI image of the patient during a procedure.
[0046] The visualizations are rendered visualizations that can combine multiple sources of data to provide visualizations of spatially encoded tool position and orientation with anatomical structure and can be used to provide position
(0044] The actual visualization can be shown on a screen or display so that the map and/or anatomical or tool structure is in a flat 2-D view and/or in 2-D what appears to be 3-D volumetric images with data representing features or electrical output with different visual characteristics such as with differing intensity, opacity, color, texture and the like. A 4-D map illustrates time-dependent activity, such as electrical activity or blood flow movement.
[0045] The systems are configured to operate based on known physical characteristics of one or more surgical tools such that the hardware is a point of interface for the circuit or software. The systems can communicate with databases that define dimensions, configurations or shapes and spacing of components on the tool(s). The defined physical data can be obtained from a CAD model of a tool.
The physical characteristics can include dimensions or other physical features or attributes and may also include relative changes in position of certain components or features upon a change in position of a tool or portion thereof. The defined physical characteristics can be electronically (programmatically) accessible by the system or known a priori and electronically stored locally or remotely and used to automatically calculate certain information and/or to segment image data. That is, the tool data from the model can be used to segment image data and/or correlate a position and orientation of a tool and/or provide trajectory adjustment guidelines or error estimates, warnings of improper trajectories and the like. For example, a grid for marking a burr hole location and/or a trajectory guide that adjusts an intrabrain path for placing a diagnostic or therapy device and such can be input, transposed, and/or overlayed in a visualization of the tool and patient structure or otherwise used, such as, for example, to project the information onto a patient's anatomical structure or determine certain operational parameters including which image volume to obtain high resolution MRI
image data that include select portions of the targeting canula. At least some of the resulting visualizations are not merely an MRI image of the patient during a procedure.
[0046] The visualizations are rendered visualizations that can combine multiple sources of data to provide visualizations of spatially encoded tool position and orientation with anatomical structure and can be used to provide position
8 adjustment data output so that a clinician can move a controller a certain amount to obtain a desired trajectory path, thereby providing a smart-adjustment system without requiring undue "guess" work on what adjustments to make to obtain the desired traj ectory.
[0047] The term "animation" refers to a sequence or series of images shown in succession, typically in relatively quick succession, such as in about 1-50 frames per second. The term "frame" refers to a single visualization or static image.
The term "animation frame" refers to one image frame of the different images in the sequence of images. The term "ACPC coordinate space" refers to a right-handed coordinate system defined by anterior and posterior commissures (AC, PC) and Mid-Sagittal plane points, with positive directions corresponding to a patient's anatomical Right, Anterior and Head directions with origin at the mid-comissure point.
[0048] The term "grid" refers to a pattern of crossed lines or shapes used as a reference for locating points or small spaces, e.g., a series of rows and intersecting columns, such as horizontal rows and vertical columns (but orientations other than vertical and horizontal can also be used). The grid can include associated visual indicia such as alphabetical markings (e.g., A-Z and the like) for rows and numbers for columns (e.g., 1-10) or the reverse. Other marking indicia may also be used. The grid can be provided as a flexible patch that can be releasably attached to the skull of a patient. For additional description of suitable grid devices, see co-pending, co-assigned U.S. Patent Application Serial No. 12/236,621.
[0049] The term "fiducial marker" refers to a marker that can be electronically identified using image recognition and/or electronic interrogation of MRI image data. The fiducial marker can be provided in any suitable manner, such as, but not limited to, a geometric shape of a portion of the tool, a component on or in the tool, a coating or fluid-filled component or feature (or combinations of different types of fiducial markers) that makes the fiducial marker(s) MRI-visible with sufficient signal intensity (brightness) for identifying location and/or orientation information for the tool and/or components thereof in space.
[0050] The term "RF safe" means that the lead or probe is configured to safely operate when exposed to RF signals, particularly RF signals associated with MRI systems, without inducing unplanned current that inadvertently unduly heats local tissue or interferes with the planned therapy. The term "MR] visible"
means that
[0047] The term "animation" refers to a sequence or series of images shown in succession, typically in relatively quick succession, such as in about 1-50 frames per second. The term "frame" refers to a single visualization or static image.
The term "animation frame" refers to one image frame of the different images in the sequence of images. The term "ACPC coordinate space" refers to a right-handed coordinate system defined by anterior and posterior commissures (AC, PC) and Mid-Sagittal plane points, with positive directions corresponding to a patient's anatomical Right, Anterior and Head directions with origin at the mid-comissure point.
[0048] The term "grid" refers to a pattern of crossed lines or shapes used as a reference for locating points or small spaces, e.g., a series of rows and intersecting columns, such as horizontal rows and vertical columns (but orientations other than vertical and horizontal can also be used). The grid can include associated visual indicia such as alphabetical markings (e.g., A-Z and the like) for rows and numbers for columns (e.g., 1-10) or the reverse. Other marking indicia may also be used. The grid can be provided as a flexible patch that can be releasably attached to the skull of a patient. For additional description of suitable grid devices, see co-pending, co-assigned U.S. Patent Application Serial No. 12/236,621.
[0049] The term "fiducial marker" refers to a marker that can be electronically identified using image recognition and/or electronic interrogation of MRI image data. The fiducial marker can be provided in any suitable manner, such as, but not limited to, a geometric shape of a portion of the tool, a component on or in the tool, a coating or fluid-filled component or feature (or combinations of different types of fiducial markers) that makes the fiducial marker(s) MRI-visible with sufficient signal intensity (brightness) for identifying location and/or orientation information for the tool and/or components thereof in space.
[0050] The term "RF safe" means that the lead or probe is configured to safely operate when exposed to RF signals, particularly RF signals associated with MRI systems, without inducing unplanned current that inadvertently unduly heats local tissue or interferes with the planned therapy. The term "MR] visible"
means that
9 the device is visible, directly or indirectly, in an MRI image. The visibility may be indicated by the increased SNR of the MRI signal proximate the device.
[0051] The system can include an intrabody MRI receive imaging probe antenna to collect signal from local tissue. The term "MRI compatible" means that the so-called component(s) is safe for use in an MRI environment and as such is typically made of a non-ferromagnetic MRI compatible material(s) suitable to reside and/or operate in a high magnetic field environment. The term "high-magnetic field"
refers to field strengths above about 0.5 T, typically above 1.0T, and more typically between about 1.5T and 10T. MRI Scanners are well known and include high-field closed bore and open bore systems.
[0052] Embodiments of the present invention can be configured to carry out diagnostic and interventional procedures such as to guide and/or place interventional devices to any desired internal region of the body or object, but may be particularly suitable for neurosurgeries. The object can be any object, and may be particularly suitable for animal and/or human subjects. Although primarily described with respect to placement of stimulation leads in the brain, the invention is not limited thereto. For example, the system can be used for gene and/or stem-cell based therapy delivery or other neural therapy delivery and allow user-defined custom targets in the brain or to other locations. In addition, embodiments of the systems can be used to ablate tissue in the brain or other locations. In some embodiments, it is contemplated that the systems can be configured to treat AFIB in cardiac tissue, and/or to deliver stem cells or other cardio-rebuilding cells or products into cardiac tissue, such as a heart wall, via a minimally invasive MRI guided procedure while the heart is beating (i.e., not requiring a non-beating heart with the patient on a heart-lung machine).
[0053] Examples of known treatments and/or target body regions are described in U.S. Patent Nos. 6,708,064; 6,438,423; 6,356,786; 6,526,318;
6,405,079;
6,167,311; 6539,263; 6,609,030 and 6,050,992, the contents of which are hereby incorporated by reference as if recited in full herein.
[0054] Embodiments of the present invention may take the form of an entirely software embodiment or an embodiment combining software and hardware aspects, all generally referred to herein as a"circuit" or "module." In some embodiments, the circuits include both software and hardware and the software is configured to work with specific hardware with known physical attributes and/or configurations. Furthermore, the present invention may take the form of a computer program product on a computer-usable storage medium having computer-usable program code embodied in the medium. Any suitable computer readable medium may be utilized including hard disks, CD-ROMs, optical storage devices, a transmission media such as those supporting the Internet or an intranet, or other storage devices.
[0055] Computer program code for carrying out operations of the present invention may be written in an object oriented programming language such as Java , Smalltalk or C++. However, the computer program code for carrying out operations of the present invention may also be written in conventional procedural programming languages, such as the "C" programming language. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on another computer, local and/or remote or entirely on the other local or remote computer. In the latter scenario, the other local or remote computer may be connected to the user's computer through a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
[0056] The present invention is described in part below with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
[0057] These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function/act specified in the flowchart and/or block diagram block or blocks.
100581 The computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
[0059] The flowcharts and block diagrams of certain of the figures herein illustrate exemplary architecture, functionality, and operation of possible implementations of embodiments of the present invention. In this regard, each block in the flow charts or block diagrams represents a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that in some alternative implementations, the functions noted in the blocks may occur out of the order noted in the figures. For example, two blocks shown in succession may in fact be executed substantially concurrently or the blocks may sometimes be executed in the reverse order or two or more blocks may be combined, depending upon the functionality involved.
[0060] Generally stated, embodiments of the systems are configured to provide a substantially automated or semi-automated and relatively easy-to-use MRI-guided systems with defined workflow steps and interactive visualizations. In particular embodiments, the systems define and present workflow with discrete steps for finding target and entry point(s), localizing the entry point(s) to a physical identified grid position, guiding the alignment of the targeting canula to a planned trajectory, monitoring the insertion of the probe, and adjusting the X-Y
position in cases where the placement needs to be corrected. During steps where specific MR
scans are used, the circuit or computer module can display data for scan plane center and angulation to be entered at the console. The workstation/circuit can passively or actively communicate with the MR scanner. The system can also be configured to use functional patient data (e.g., fiber tracks, fMRI and the like) to help plan or refine a target surgical site.
[0061] Embodiments of the present invention will now be described in further detail below with reference to the figures. Figure 1 illustrates an MRI guided interventional system 10 with an MRI scanner 20, a clinician workstation 30 with at least one circuit 30c, at least one display 32 and at least one MRI compatible interventional and/or surgical tool 50. An MRI scanner interface 40 may be used to allow communication between the workstation 30 and the scanner 20. The interface 40 and/or circuit 30c may be hardware, software or a combination of same. The interface 40 and/or circuit 30c may reside partially or totally in the scanner 20, partially or totally in the workstation 30, or partially or totally in a discrete device therebetween. The system 10 can be configured to render or'generate real time visualizations of the target anatomical space using MRI image data and predefined data of at least one surgical tool to segment the image data and place the tool 50 in the rendered visualization in the correct orientation and position in 3D space, anatomically registered to a patient. The tool 50 can include or cooperate with tracking, monitoring and/or interventional components. The system 10 can optionally include a reader 30r that can electronically read (e.g., optically such as via a bar code or otherwise electronically read such a via an RFID tag) a label or tag or other indicia to confirm that the hardware is authentic or compatible to inhibit counterfeit hardware and potential misuse of the system as the system is configured so that certain hardware define a point of interface with the software or circuit 30c.
Alternatively, or additionally, the system 10 can allow a user to manually input the tool/hardware indicia. Proper operation of the system requires that the proper hardware having the specific predefined characteristics used by the system is used for the surgical procedure.
[0062] Figures 2 and 3 are schematics of embodiments of the system 10 which illustrate that the system 10 can include a light source 100 in communication with a camera/imaging device 110 via a fiber optic fiber bundle cable 115.
Figure 2 illustrates that the system 10 can be used for bilateral procedures. The camera device (with fiber optic cables) 110 can have a distal lens and can be configured with a relatively small local field of view (residing proximate the burr hole or surgical entry location) to allow a clinician to monitor the surgical entry point. The camera device 110 can be mounted to the trajectory guide 50. The return signal is fed to an MRI
compatible video camera 120 and the signal is transmitted as a video of the patient and can be shown in a display or split screen 32 at the workstation 30. The workstation 30 can be in a control room 200 and the feed from the fiber optic cable from the camera 115c can be via an RF filter 123 to inhibit signal distortion to the video stream shown on the display 32. A separate display or monitor can also reside in the surgical room 210. A sterile surgical drape 118 can be used to maintain a sterile side inside the surgical room 210 on the bore end of the magnet facing the camera 120. On example of a suitable MR compatible video camera is available from MRC Systems GmbH, Heidelberg, Germany.
[00631 The system 10 can be configured to provide workflow for a unilateral or bilateral (or even trilateral or more) procedure. Selection of the procedure type can initiate the associated work flow presented. Figure 4 illustrates an example of a workstation control panel 30p on display screen 32. The pane130p can illustrate a current workflow step and allow a user to go to a step directly (such as via a drop down list or selection of a workflow step in a toolbar or the like) and can be presented adjacent different views of the intrabody trajectory and patient anatomy.
Tabs or other user-selectable features with visual feedback on status of a step for each side can be used for steps with laterality (e.g., left or right for bilateral procedures) to allow a user to control selection of laterality, such as left 30a and right 30b, to complete trajectory planning for each side independently (or to allow a user to toggle back and forth while maintaining control over each side). The display 32 can include viewer tools such as zoom, pan, width/level, magnifier, etc.
[0064] The MRI scanner 20 can include a console that has a "launch"
application or portal for allowing communication to the circuit 30c of the workstation 30. The scanner console can acquire volumetric T1-weighted (post-contrast scan) data or other image data (e.g., high resolution image data for a specific volume) of a patient's head or other anatomy. In some embodiments, the console can push DICOM
images or other suitable image data to the workstation 30 and/or circuit 30c.
The workstation 30 and/or circuit 30c can be configured to passively wait for data to be sent from the MR scanner 20 and the circuit 30c/workstation 30 does not query the Scanner or initiate a communication to the Scanner. In other embodiments, a dynamic or active communication protocol between the circuit 30c/workstation and the Scanner 20 may be used to acquire image data and initiate or request particular scans and/or scan volumes. Also, in some embodiments, pre-DICOM, but reconstructed image data, can be sent to the circuit 30c/workstation 30 for processing or display. In other embodiments, pre-reconstruction image data (e.g., substantially "raw" image data) can be sent to the circuit 30c/workstation 30 for Fourier Transform and reconstruction.
[0065] Generally described, for some unilateral scenarios, the user will proceed through a set of discrete workflow steps to load MR image data, identify a target point, identify an entry point, verify the planned trajectory, and align the targeting canula. A target point or region can also be planned or refined based on real-time functional image data of a patient. The functional image data can include, but is not limited to, images of fiber tracks, images of activity in brain regions during vocalization (e.g., reading, singing, talking), or based on physical or olefactory or sense-based stimulation, such as exposure to electrical (discomfort/shock input), heat and/or cold, light or dark, visual images, pictures or movies, chemicals, scents, taste, and sounds or the like) and/or using fMRI or other imaging techniques. The enhanced visualization gives neurosurgeons a much clearer picture of the spatial relationship of a patient's brain structures. The visualizations can serve as a trajectory guide for surgical procedures, such as brain-tumor removal and epilepsy surgery. In some embodiments, the visualizations can be generated using data collated from different types of brain-imaging methods, including conventional magnetic resonance imaging (MRI), functional MRI (fiVIRI), diffusion-tensor imaging (DTI) and even hyperpolarized noble gas MRI imaging. The MRI gives details on the anatomy, IMRI
or other active stimulation-based imaging protocol can provide information on the activated areas of the brain, and DTI provides images of the network of nerve fibers connecting different brain areas. The fusion of one or all of these different images and the tool information can be used to produce a 3-D display with trajectory information that surgeons can manipulate.
[00661 Thus, a target location and trajectory can be planned, confirmed or refined based in-part on functional information of the patient. This functional information can be provided in real-time visualizations of the patient with the trajectory guide tools for trajectory path and target planning, e.g., visualize a patient's fiber track structures and/or functional information of a patient's brain for a surgeon's ease of reference. This information can also be selected or suppressed from views via a UI selection, such as "Show Fiber Tracks" 32FI and/or "Show Functional Output"
32F2 (e.g., toolbar option) as shown in Figure 4. It is noted that the patient functional information can be shown automatically without requiring a user selection or in response to a stage of a procedure or when selecting certain steps. In addition, such information may be shown or selected in any appropriate display or step described herein although not specifically described with respect to that particular step or screen display. Knowing where susceptible or sensitive brain regions are or where critical fiber tracks are in the patient's brain, can allow a surgeon to plan a better, less-risky or less-intrusive trajectory and/or allow a surgeon to more precisely reach a desired target site and/or more precisely place a device and/or deliver a planned therapy, e.g., implant a stimulation lead, ablate tissue and/or treat a tumor site and/or excise a tumor, deliver a gene and/or stem cell therapy and the like.
[00671 To align the targeting canula, scan volumes can be defined by the system based on known dimensions of the canula, such as a canula length a position of a proximal or distal marker on the canula, and angulation and lateral (X-Y) pivot limit. In particular embodiments, the user can then gradually advance a probe and a peel away sheath (that is configured to guide an interventional device to a desired location along the defined trajectory) and acquire images to check for hemorrhage and to verify the trajectory and/or avoid functionally sensitive structure. When the probe has been advanced to the target point, high-resolution confirmation images can be obtained to verify the tip location relative to the planned location. If actual placement is unacceptable, the probe can be withdrawn. At that point, either the X-Y
placement can be adjusted appropriately (e.g., by moving a platform or stage an amount to cause the desired adjustment) or a trajectory angulation can be re-planned and a second attempt can be made.
[0068) For some bilateral scenarios, the above steps can be repeated for both left and right sides, with the additional goal that the patient should not be moved into or out of the scanner. To satisfy that goal, trajectory planning should be completed for both sides prior to removing the patient from the scanner. Also, burring and frame attachment (the member that holds the trajectory guide to the patient's head) should be completed for both sides prior to moving the patient back into the scanner to promote speed of the procedure.
[0069] The system 10 can be configured with a hardware interface that provides a network connection, e.g., a standard TCP/IP over Ethernet network connection, to provide access to MR scanner 20, such as the DICOM server. The workstation 30 can provide a DICOM C-STORE storage class provider. The scanner console can be configured to be able to push images to the workstation 30 and the workstation 30 can be configured to directly or indirectly receive DICOM MR
image data pushed from an MR Scanner console. Altematively, as noted above, the system can be configured with an interface that allows for a dynamic and interactive communication with the Scanner 20 and can obtain image data in other formats and stages (e.g., pre-DICOM reconstructed or raw image data).
[0070] As noted above, the systems 10 are configured so that hardware, e.g., one or more specific surgical tools, constitute a point of interface with the system (software or computer programs) because the circuit 30c is configured with predefined tool data that recognizes physical characteristics of specific tool hardware.
[0071] As shown in Figure 5, to assure proper operation, the system 10 can be configured to require entry of a valid identifier and/or revision controlled/based part number to validate that the hardware planned for use is appropriate for use in the system 10 (or at least that version of the system).
Thus, a reader 30r associated with the workstation 30 can be configured to read a single "group" identifier 66 that can be placed on the kit package or provided with the kit 10k and/or the reader can be configured to read each tool that has predefined characteristics to confirm the appropriate part and version is in the kit.
Altematively, or additionally, the system 10 can allow a user to manually input the tool/hardware identifier data (e.g., hardware version and/or part number) into a UI
associated with the circuit 30c. The workstation 30 can include a look-up chart of a correlation table 31 that confirms the correct hardware is in the kit 10k or otherwise provided for use.
Thus, the workstation 30 can be configured with a user interface 321 (shown as a Tool Version Identifier Panel) that requires a user to electronically or manually enter the identifier 66 and/or to acknowledge compliance with the tool-specific operation of the system 10. It is envisioned that the circuit 30c can be configured with updates and backward compatibility for future controlled changes to the specific tools and/or with the ability to use different module versions of the system according to the version of the tool or tools then in use at the surgical site.
(0072] To inhibit the use of counterfeit hardware with the system 10, the identifier may include indicia that can be keyed to a particular authorized use site and/or authorized user. The system may be configured to require a user to certify that the hardware is OEM hardware or authorized hardware to be able to receive an electronic key to be able to activate the system. A user may be required to contact the OEM or other authorized party to obtain an electronic key or identifier to allow use of the hardware with the system 10.
[0073] As shown in Figure 5, in some embodiments, the system is programmatically configured to recognize defined physical characteristics of different tools. Those tools that can be provided as a kit 10k (typically a single-use disposable hardware) or in other groups or sub-groups or even individually, typically provided in suitable sterile packaging. The tools can include at least one marking grid 50g (also referred to as a grid patch), a targeting canula 60 with a distal marker 60m and an opposing proximal portion of the canula 60p. The targeting canula 60 can include an open center lumen or passage 61 (Figure 8). The distal maker 60m typically includes a substantially spherical fluid filled component 65 (Figure 8). The proximal portion of the canula 60p can include a marker, but is typically identified in the image data based, at least in-part) on the distal maker 60m and its known distance and orientation with respect thereto based on the physical characteristics of the targeting canula 60.
Still referring to Figure 5, the system 10 can also include a trajectory guide 50t with a plurality of MRI visible frame fiducial markers 50fm around a base 50b thereof. The system 10 may also include a stylet that can communicate with a peel-away sheath 50s and an imaging probe 50a (that provides an intrabody receive antenna that can be slidably introduced via the passage of the targeting canula 60). Certain components of the kit may be replaced or omitted depending on the desired procedure.
Certain components can be provided in duplicate for bilateral procedures. As shown in Figure 5, a therapy delivery device 52 may optionally be provided, also with an identifier such as a label or tag 66. The device 52 can be configured to flowably introduce and/or inject a desired therapy (e.g., gene therapy or stem-cell or other therapy type) or obtain a biopsy and the like.
[0074] Figures 6A-6C, and 7-10 illustrate the trajectory guide 50t with the targeting canula 60 and various features described above. Figure 6A
illustrates a trajectory guide 50t and targeting canula 60 in position on a patient with trajectory guide actuators 151 and respective actuator cables 150a-150d (providing X-Y
adjustment and pitch and roll adjustment) in communication with a trajectory adjustment controller 400. The frame 150f can include control arcs 152 (Figures 7,
[0051] The system can include an intrabody MRI receive imaging probe antenna to collect signal from local tissue. The term "MRI compatible" means that the so-called component(s) is safe for use in an MRI environment and as such is typically made of a non-ferromagnetic MRI compatible material(s) suitable to reside and/or operate in a high magnetic field environment. The term "high-magnetic field"
refers to field strengths above about 0.5 T, typically above 1.0T, and more typically between about 1.5T and 10T. MRI Scanners are well known and include high-field closed bore and open bore systems.
[0052] Embodiments of the present invention can be configured to carry out diagnostic and interventional procedures such as to guide and/or place interventional devices to any desired internal region of the body or object, but may be particularly suitable for neurosurgeries. The object can be any object, and may be particularly suitable for animal and/or human subjects. Although primarily described with respect to placement of stimulation leads in the brain, the invention is not limited thereto. For example, the system can be used for gene and/or stem-cell based therapy delivery or other neural therapy delivery and allow user-defined custom targets in the brain or to other locations. In addition, embodiments of the systems can be used to ablate tissue in the brain or other locations. In some embodiments, it is contemplated that the systems can be configured to treat AFIB in cardiac tissue, and/or to deliver stem cells or other cardio-rebuilding cells or products into cardiac tissue, such as a heart wall, via a minimally invasive MRI guided procedure while the heart is beating (i.e., not requiring a non-beating heart with the patient on a heart-lung machine).
[0053] Examples of known treatments and/or target body regions are described in U.S. Patent Nos. 6,708,064; 6,438,423; 6,356,786; 6,526,318;
6,405,079;
6,167,311; 6539,263; 6,609,030 and 6,050,992, the contents of which are hereby incorporated by reference as if recited in full herein.
[0054] Embodiments of the present invention may take the form of an entirely software embodiment or an embodiment combining software and hardware aspects, all generally referred to herein as a"circuit" or "module." In some embodiments, the circuits include both software and hardware and the software is configured to work with specific hardware with known physical attributes and/or configurations. Furthermore, the present invention may take the form of a computer program product on a computer-usable storage medium having computer-usable program code embodied in the medium. Any suitable computer readable medium may be utilized including hard disks, CD-ROMs, optical storage devices, a transmission media such as those supporting the Internet or an intranet, or other storage devices.
[0055] Computer program code for carrying out operations of the present invention may be written in an object oriented programming language such as Java , Smalltalk or C++. However, the computer program code for carrying out operations of the present invention may also be written in conventional procedural programming languages, such as the "C" programming language. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on another computer, local and/or remote or entirely on the other local or remote computer. In the latter scenario, the other local or remote computer may be connected to the user's computer through a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
[0056] The present invention is described in part below with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
[0057] These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function/act specified in the flowchart and/or block diagram block or blocks.
100581 The computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
[0059] The flowcharts and block diagrams of certain of the figures herein illustrate exemplary architecture, functionality, and operation of possible implementations of embodiments of the present invention. In this regard, each block in the flow charts or block diagrams represents a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that in some alternative implementations, the functions noted in the blocks may occur out of the order noted in the figures. For example, two blocks shown in succession may in fact be executed substantially concurrently or the blocks may sometimes be executed in the reverse order or two or more blocks may be combined, depending upon the functionality involved.
[0060] Generally stated, embodiments of the systems are configured to provide a substantially automated or semi-automated and relatively easy-to-use MRI-guided systems with defined workflow steps and interactive visualizations. In particular embodiments, the systems define and present workflow with discrete steps for finding target and entry point(s), localizing the entry point(s) to a physical identified grid position, guiding the alignment of the targeting canula to a planned trajectory, monitoring the insertion of the probe, and adjusting the X-Y
position in cases where the placement needs to be corrected. During steps where specific MR
scans are used, the circuit or computer module can display data for scan plane center and angulation to be entered at the console. The workstation/circuit can passively or actively communicate with the MR scanner. The system can also be configured to use functional patient data (e.g., fiber tracks, fMRI and the like) to help plan or refine a target surgical site.
[0061] Embodiments of the present invention will now be described in further detail below with reference to the figures. Figure 1 illustrates an MRI guided interventional system 10 with an MRI scanner 20, a clinician workstation 30 with at least one circuit 30c, at least one display 32 and at least one MRI compatible interventional and/or surgical tool 50. An MRI scanner interface 40 may be used to allow communication between the workstation 30 and the scanner 20. The interface 40 and/or circuit 30c may be hardware, software or a combination of same. The interface 40 and/or circuit 30c may reside partially or totally in the scanner 20, partially or totally in the workstation 30, or partially or totally in a discrete device therebetween. The system 10 can be configured to render or'generate real time visualizations of the target anatomical space using MRI image data and predefined data of at least one surgical tool to segment the image data and place the tool 50 in the rendered visualization in the correct orientation and position in 3D space, anatomically registered to a patient. The tool 50 can include or cooperate with tracking, monitoring and/or interventional components. The system 10 can optionally include a reader 30r that can electronically read (e.g., optically such as via a bar code or otherwise electronically read such a via an RFID tag) a label or tag or other indicia to confirm that the hardware is authentic or compatible to inhibit counterfeit hardware and potential misuse of the system as the system is configured so that certain hardware define a point of interface with the software or circuit 30c.
Alternatively, or additionally, the system 10 can allow a user to manually input the tool/hardware indicia. Proper operation of the system requires that the proper hardware having the specific predefined characteristics used by the system is used for the surgical procedure.
[0062] Figures 2 and 3 are schematics of embodiments of the system 10 which illustrate that the system 10 can include a light source 100 in communication with a camera/imaging device 110 via a fiber optic fiber bundle cable 115.
Figure 2 illustrates that the system 10 can be used for bilateral procedures. The camera device (with fiber optic cables) 110 can have a distal lens and can be configured with a relatively small local field of view (residing proximate the burr hole or surgical entry location) to allow a clinician to monitor the surgical entry point. The camera device 110 can be mounted to the trajectory guide 50. The return signal is fed to an MRI
compatible video camera 120 and the signal is transmitted as a video of the patient and can be shown in a display or split screen 32 at the workstation 30. The workstation 30 can be in a control room 200 and the feed from the fiber optic cable from the camera 115c can be via an RF filter 123 to inhibit signal distortion to the video stream shown on the display 32. A separate display or monitor can also reside in the surgical room 210. A sterile surgical drape 118 can be used to maintain a sterile side inside the surgical room 210 on the bore end of the magnet facing the camera 120. On example of a suitable MR compatible video camera is available from MRC Systems GmbH, Heidelberg, Germany.
[00631 The system 10 can be configured to provide workflow for a unilateral or bilateral (or even trilateral or more) procedure. Selection of the procedure type can initiate the associated work flow presented. Figure 4 illustrates an example of a workstation control panel 30p on display screen 32. The pane130p can illustrate a current workflow step and allow a user to go to a step directly (such as via a drop down list or selection of a workflow step in a toolbar or the like) and can be presented adjacent different views of the intrabody trajectory and patient anatomy.
Tabs or other user-selectable features with visual feedback on status of a step for each side can be used for steps with laterality (e.g., left or right for bilateral procedures) to allow a user to control selection of laterality, such as left 30a and right 30b, to complete trajectory planning for each side independently (or to allow a user to toggle back and forth while maintaining control over each side). The display 32 can include viewer tools such as zoom, pan, width/level, magnifier, etc.
[0064] The MRI scanner 20 can include a console that has a "launch"
application or portal for allowing communication to the circuit 30c of the workstation 30. The scanner console can acquire volumetric T1-weighted (post-contrast scan) data or other image data (e.g., high resolution image data for a specific volume) of a patient's head or other anatomy. In some embodiments, the console can push DICOM
images or other suitable image data to the workstation 30 and/or circuit 30c.
The workstation 30 and/or circuit 30c can be configured to passively wait for data to be sent from the MR scanner 20 and the circuit 30c/workstation 30 does not query the Scanner or initiate a communication to the Scanner. In other embodiments, a dynamic or active communication protocol between the circuit 30c/workstation and the Scanner 20 may be used to acquire image data and initiate or request particular scans and/or scan volumes. Also, in some embodiments, pre-DICOM, but reconstructed image data, can be sent to the circuit 30c/workstation 30 for processing or display. In other embodiments, pre-reconstruction image data (e.g., substantially "raw" image data) can be sent to the circuit 30c/workstation 30 for Fourier Transform and reconstruction.
[0065] Generally described, for some unilateral scenarios, the user will proceed through a set of discrete workflow steps to load MR image data, identify a target point, identify an entry point, verify the planned trajectory, and align the targeting canula. A target point or region can also be planned or refined based on real-time functional image data of a patient. The functional image data can include, but is not limited to, images of fiber tracks, images of activity in brain regions during vocalization (e.g., reading, singing, talking), or based on physical or olefactory or sense-based stimulation, such as exposure to electrical (discomfort/shock input), heat and/or cold, light or dark, visual images, pictures or movies, chemicals, scents, taste, and sounds or the like) and/or using fMRI or other imaging techniques. The enhanced visualization gives neurosurgeons a much clearer picture of the spatial relationship of a patient's brain structures. The visualizations can serve as a trajectory guide for surgical procedures, such as brain-tumor removal and epilepsy surgery. In some embodiments, the visualizations can be generated using data collated from different types of brain-imaging methods, including conventional magnetic resonance imaging (MRI), functional MRI (fiVIRI), diffusion-tensor imaging (DTI) and even hyperpolarized noble gas MRI imaging. The MRI gives details on the anatomy, IMRI
or other active stimulation-based imaging protocol can provide information on the activated areas of the brain, and DTI provides images of the network of nerve fibers connecting different brain areas. The fusion of one or all of these different images and the tool information can be used to produce a 3-D display with trajectory information that surgeons can manipulate.
[00661 Thus, a target location and trajectory can be planned, confirmed or refined based in-part on functional information of the patient. This functional information can be provided in real-time visualizations of the patient with the trajectory guide tools for trajectory path and target planning, e.g., visualize a patient's fiber track structures and/or functional information of a patient's brain for a surgeon's ease of reference. This information can also be selected or suppressed from views via a UI selection, such as "Show Fiber Tracks" 32FI and/or "Show Functional Output"
32F2 (e.g., toolbar option) as shown in Figure 4. It is noted that the patient functional information can be shown automatically without requiring a user selection or in response to a stage of a procedure or when selecting certain steps. In addition, such information may be shown or selected in any appropriate display or step described herein although not specifically described with respect to that particular step or screen display. Knowing where susceptible or sensitive brain regions are or where critical fiber tracks are in the patient's brain, can allow a surgeon to plan a better, less-risky or less-intrusive trajectory and/or allow a surgeon to more precisely reach a desired target site and/or more precisely place a device and/or deliver a planned therapy, e.g., implant a stimulation lead, ablate tissue and/or treat a tumor site and/or excise a tumor, deliver a gene and/or stem cell therapy and the like.
[00671 To align the targeting canula, scan volumes can be defined by the system based on known dimensions of the canula, such as a canula length a position of a proximal or distal marker on the canula, and angulation and lateral (X-Y) pivot limit. In particular embodiments, the user can then gradually advance a probe and a peel away sheath (that is configured to guide an interventional device to a desired location along the defined trajectory) and acquire images to check for hemorrhage and to verify the trajectory and/or avoid functionally sensitive structure. When the probe has been advanced to the target point, high-resolution confirmation images can be obtained to verify the tip location relative to the planned location. If actual placement is unacceptable, the probe can be withdrawn. At that point, either the X-Y
placement can be adjusted appropriately (e.g., by moving a platform or stage an amount to cause the desired adjustment) or a trajectory angulation can be re-planned and a second attempt can be made.
[0068) For some bilateral scenarios, the above steps can be repeated for both left and right sides, with the additional goal that the patient should not be moved into or out of the scanner. To satisfy that goal, trajectory planning should be completed for both sides prior to removing the patient from the scanner. Also, burring and frame attachment (the member that holds the trajectory guide to the patient's head) should be completed for both sides prior to moving the patient back into the scanner to promote speed of the procedure.
[0069] The system 10 can be configured with a hardware interface that provides a network connection, e.g., a standard TCP/IP over Ethernet network connection, to provide access to MR scanner 20, such as the DICOM server. The workstation 30 can provide a DICOM C-STORE storage class provider. The scanner console can be configured to be able to push images to the workstation 30 and the workstation 30 can be configured to directly or indirectly receive DICOM MR
image data pushed from an MR Scanner console. Altematively, as noted above, the system can be configured with an interface that allows for a dynamic and interactive communication with the Scanner 20 and can obtain image data in other formats and stages (e.g., pre-DICOM reconstructed or raw image data).
[0070] As noted above, the systems 10 are configured so that hardware, e.g., one or more specific surgical tools, constitute a point of interface with the system (software or computer programs) because the circuit 30c is configured with predefined tool data that recognizes physical characteristics of specific tool hardware.
[0071] As shown in Figure 5, to assure proper operation, the system 10 can be configured to require entry of a valid identifier and/or revision controlled/based part number to validate that the hardware planned for use is appropriate for use in the system 10 (or at least that version of the system).
Thus, a reader 30r associated with the workstation 30 can be configured to read a single "group" identifier 66 that can be placed on the kit package or provided with the kit 10k and/or the reader can be configured to read each tool that has predefined characteristics to confirm the appropriate part and version is in the kit.
Altematively, or additionally, the system 10 can allow a user to manually input the tool/hardware identifier data (e.g., hardware version and/or part number) into a UI
associated with the circuit 30c. The workstation 30 can include a look-up chart of a correlation table 31 that confirms the correct hardware is in the kit 10k or otherwise provided for use.
Thus, the workstation 30 can be configured with a user interface 321 (shown as a Tool Version Identifier Panel) that requires a user to electronically or manually enter the identifier 66 and/or to acknowledge compliance with the tool-specific operation of the system 10. It is envisioned that the circuit 30c can be configured with updates and backward compatibility for future controlled changes to the specific tools and/or with the ability to use different module versions of the system according to the version of the tool or tools then in use at the surgical site.
(0072] To inhibit the use of counterfeit hardware with the system 10, the identifier may include indicia that can be keyed to a particular authorized use site and/or authorized user. The system may be configured to require a user to certify that the hardware is OEM hardware or authorized hardware to be able to receive an electronic key to be able to activate the system. A user may be required to contact the OEM or other authorized party to obtain an electronic key or identifier to allow use of the hardware with the system 10.
[0073] As shown in Figure 5, in some embodiments, the system is programmatically configured to recognize defined physical characteristics of different tools. Those tools that can be provided as a kit 10k (typically a single-use disposable hardware) or in other groups or sub-groups or even individually, typically provided in suitable sterile packaging. The tools can include at least one marking grid 50g (also referred to as a grid patch), a targeting canula 60 with a distal marker 60m and an opposing proximal portion of the canula 60p. The targeting canula 60 can include an open center lumen or passage 61 (Figure 8). The distal maker 60m typically includes a substantially spherical fluid filled component 65 (Figure 8). The proximal portion of the canula 60p can include a marker, but is typically identified in the image data based, at least in-part) on the distal maker 60m and its known distance and orientation with respect thereto based on the physical characteristics of the targeting canula 60.
Still referring to Figure 5, the system 10 can also include a trajectory guide 50t with a plurality of MRI visible frame fiducial markers 50fm around a base 50b thereof. The system 10 may also include a stylet that can communicate with a peel-away sheath 50s and an imaging probe 50a (that provides an intrabody receive antenna that can be slidably introduced via the passage of the targeting canula 60). Certain components of the kit may be replaced or omitted depending on the desired procedure.
Certain components can be provided in duplicate for bilateral procedures. As shown in Figure 5, a therapy delivery device 52 may optionally be provided, also with an identifier such as a label or tag 66. The device 52 can be configured to flowably introduce and/or inject a desired therapy (e.g., gene therapy or stem-cell or other therapy type) or obtain a biopsy and the like.
[0074] Figures 6A-6C, and 7-10 illustrate the trajectory guide 50t with the targeting canula 60 and various features described above. Figure 6A
illustrates a trajectory guide 50t and targeting canula 60 in position on a patient with trajectory guide actuators 151 and respective actuator cables 150a-150d (providing X-Y
adjustment and pitch and roll adjustment) in communication with a trajectory adjustment controller 400. The frame 150f can include control arcs 152 (Figures 7,
10) that cooperate with a platform 153 (Figure 7) to provide pitch and roll adjustments. The platform 153 can allow for X-Y adjustments of the trajectory.
For additional discussion of suitable trajectory guides, see, U.S. Application Serial No.
12/134,412, and co-pending, co-assigned U.S. Patent Application Serial No.
identified by Attorney Docket No. 9450-341P, the contents of which are hereby incorporated by reference as if recited in full herein.
[0075] Figures 6B and 6C illustrate examples of a depth stop 210 cooperating with an elongate member 212 such as, for example, a stimulation or ablation lead, a diagnostic and/or imaging probe, a removable sheath and/or other therapy or diagnostic device inserted and secured therein as illustrated. The illustrated depth stop 210 has a generally cylindrical configuration with opposite proximal and distal ends 210a, 210b and is adapted to be removably secured within the proximal end of the tubular trajectory guide member 50t (Figure 6A). The depth stop 210 is configured to limit a distance that the member 212 extends into the body of a patient when the depth stop is inserted within the tubular member 50t or 60. The member 212 can include visual indicia of insertion depth 215 to allow a user to visually attach the stop 210 at the appropriate location that provides the desired insertion depth.
[0076] As shown in Figures 6C and 6D, in some embodiments, the depth stop 210 is attached to a peel-away sheath 212s (see also probe 50s, Figure 5) and can be configured to receive and guide an elongated interventional device, such as a stylet or imaging probe, therethrough. As shown in Figures 6C and 6D, the sheath 212s can include opposing tabs that, when pulled apart, cause the sheath to peel away for removal from the targeting canula 60. In other embodiments, the depth stop 210 can be attached to a stimulation lead to allow for a defined insertion depth (see, e.g., Figure 38 below) or other device where insertion depth control is desired.
[0077] As shown in Figure 7, the targeting canula 60 can be attached to the trajectory guide 50t. The targeting canula 60 can include a through passage 61.
The distal end of the targeting canula 60 has a fiducial marker 60m, shown as a substantially spherical or round (cross-section) marker shape. The proximal end 60p can be configured with a fluid filled channel 68 concentric with the passage 61 that can define a cylindrical fiducial marker. Along the axis of the canula 60, there is the lumen or passage 61 through which another device can be slidably introduced and/or withdrawn, e.g., a stylet 50s, imaging probe 50a, therapy delivery or diagnostic device 52 (Figure 5) and DBS stimulation leads for implantation, can be advanced into the brain.
[0078] Figure 11A illustrates an example of a grid patch 50g with physical characteristics that are predefined and available to or in the circuit 30c (e.g., software application). Figure 11B illustrates that the grid can allow for precise correlation of logical points in an MR volume with the physical patient. The system 10 can be configured to generate a 3-D volumetric view on the display 32 with overlays to show the grid 50g and an entry point and grid coordinates (and optionally grid edges) allowing a burr hole mark to be made on a skull at the desired planned entry point.
[0079] Figures 12A-12E illustrate a series of operations that can be carried out for the grid 50g segmentation. Some or all of these operations may be carried out behind the scene (e.g., not actually displayed). Figure 12A
illustrates an example of a histogram. Figure 12B is an example of an initial distance image.
Figure 12C is an image of a result of searching for the grid in the initial distance image. Figure 12D illustrates the result of searching for the grid in the optimal distance image, with small step sizes. Figure 12E illustrates the result of spatially deforming the grid to fit the head surface and interpolating the grid cells to find the vertices.
[0080] With reference to Figure 12A, the amplitude of the background noise in the input image stack can be estimated. To do so, a histogram of the stack can be constructed. The first negative maximum of the slope of the histogram can be located. The first peak to the left of this maximum can be located (this can be termed the "noise peak"). The difference between the noise peak and the first negative maximum of the slope can approximate the standard deviation of the noise. The noise threshold can be obtained using the following Equation (1).
(noise threshold) = (noise peak) + 4*(noise standard deviation) EQUATION (1) [0081] The above can be considered as a first step in the grid segmentation of the image data. Steps 2-5 can be carried out as described below to place the grid in position and deform to curvature of the skull for the grid segmentation.
2. Estimate the "optimal view direction" for the grid. This is defined as the vector from the grid center to the midpoint of the AC-PC line:
a. Construct an initial distance image as shown in Figure 12B starting from a rough estimate of the optimal view direction:
i. Create an image plane that is perpendicular to the estimated view direction, and outside of the head volume.
ii. For each pixel in the image plane, record its distance to the first voxel in the image volume that is above the noise threshold.
b. Find the grid cells in the distance image:
i. The grid cells are "bumps" on the head surface, so they appear as local minima in the distance image. This characteristic is enhanced by applying a Laplacian-of-Gaussian operator to the distance image.
ii. For each location in the image, place a virtual grid (whose dimensions match those of the physical grid), at several orientations.
Compute a score by tallying 1 point for each virtual grid cell that lands on one of the image minima. The highest score corresponds to the location and orientation of the grid in the distance image.
c. The center of the grid so obtained gives a new estimate of the optimal view direction. This estimate is used as a starting point of a new iteration of steps 2a, and 2b, in order to refine the estimate.
3. Find the grid using the optimal view direction:
a. Construct a distance-image as described in step 2a, but using the optimal view direction obtained in step 2.
b. Find the grid in the new distance image, as described in 2b.
c. Refine the grid position and orientation by repeating the search procedure of 2b, but search only in a small region around the known approximate location, with very fine step-sizes (see, Figure 12D).
4. Deform the grid to the shape of the head:
a. Up to this point, the grid has been taken to be planar.
b. Fit the grid cell locations in the distance image to a cubic surface using a robust regression algorithm, and snap each grid cell from step 3 to this surface. (This is to avoid producing a very irregular grid surface due to noise and low resolution in the distance image.) Note that, in general, it is not possible to maintain equal distances between the grid cells when they are snapped to an arbitrary curved surface.
Therefore an algorithm which simulates the behavior of the physical grid patch is used to minimize the amount of stretching and bending during the deformation.
5. Interpolate between the grid cells to compute the grid vertices. Figure 12E
illustrates a deformation result.
[00821 Figures 13A-13C illustrate images associated with a series of operations that can be carried out for frame marker segmentation. Figures 9 and 10 illustrate an example of a frame 50f with fiducial markers 50fm circumferentially spaced apart around a center "cp" and around a space for the burr hole with two of the markers 50fmt, 50fm2 closer together than a third 50fm3 (Figure 10). Different arrangements, configurations or numbers of fiducial makers as well as different locations on the frame 50f and/or guide 50t may be used and the frame segmentation can be altered accordingly.
[00831 Some or all of these operations illustrated in or described with respect to Figures 13A-C may be carried out behind the scene (e.g., not actually displayed). Figure 13A illustrates an exemplary histogram with a noise region.
Figure 13B illustrates a cross-section of an imaged frame-marker. Figure 13C
illustrates a result of fitting a circle to the edge mask shown in Figure 13B.
The frame marker segmentation can define the orientation of the trajectory guide on the patient that can be correlated to mechanical output to cause the trajectory guide 50t to translate to provide a desired trajectory, e.g., rotations and/or direction left, right, counterclockwise or clockwise or other translation for moving the actuators to change pitch and/or roll (or X-Y location).
[0084] With reference to Figure 13A, the amplitude of the background noise in the input image stack can be estimated. To do so, a histogram of the stack can be constructed, the first negative maximum of the slope of the histogram can be located, the first peak to the left of this maximum can be located (this is the noise peak). The difference between the noise peak and the first negative maximum of the slope can approximate the standard deviation of the noise. The noise threshold can be obtained using the Equation (1) above. Again, this can be considered a first step in segmentation. Then, Steps 2-6 can be carried out with respect to the frame marker segmentation based on fit of an expected fiducial geometry to the observed fiducial positions because the fiducial markers are arranged with a fixed geometric relationship inside the volume.
2. Use a region-growing algorithm to find all "clumps" of pixels that are above the noise threshold.
3. Discard all clumps whose volumes are far from the known frame-marker volume.
4. Discard all clumps whose bounding-box dimension are far from the known bounding-box dimensions of the frame-markers.
5. Among the remaining clumps, look for triplets whose spatial arrangement matches the known spatial arrangement of the frame-markers of the trajectory base 50b (i.e., the centroids of the clumps should form a triangle (e.g., an isosceles triangle) of known dimensions.) [0085] At this point, if the number of clumps found for each frame (where bilateral procedures are used, there are two frames) is not exactly 3, then the segmentation is deemed to have failed.
6. For each triplet of markers, refine the marker locations:
a. Extract a 2D image of each marker by reformatting the stack onto the plane defined by the 3 marker centroids.
b. Compute the Canny edge mask of each 2D marker image.
c. Fit a circle to the Canny edge mask. The circle diameter is set equal to the diameter of the physical frame markers. The fit is performed by moving the circle until its overlap with the edge mask is maximized. The center of the fit circle is taken to be the location of the frame-marker's center.
[0086] In some embodiments, circuit 30c can be configured so that the program application can have distinct ordered workflow steps that are organized into logical groups based on major divisions in the clinical workflow as shown in Table 1.
A user may return to previous workflow steps if desired. Subsequent workflow steps may be non-interactive if requisite steps have not been completed. The major workflow groups and steps can include the following features or steps in the general workflow steps of "start", "plan entry", "plan target", "navigate", and "refine,"
ultimately leading to delivering the therapy (here placing the stimulation lead).
TABLE 1: EXEMPLARY CLINICAL WORKFLOW GROUPS/STEPS
Grou S:te ~ Deseri Start Start Set overall procedure parameters (Optionally confirm hardware com atibi(it ACPC Acquire a volume and determine AC, PC, and MSP points Plan Target Define initial target point(s) for entry planning Entry Trajectory Explore potential trajectories to determine entry oint s Grid Locate physical entry point via fiducial grid.
ACPC With hole burred and frame attached, acquire a volume Plan and determine revised AC, PC, and MSP points.
Target Target Acquire high-resolution slabs (e.g., T2 slabs) to determine target positions in new volume.
Trajectory Review final planned trajectory prior to starting procedure.
Initiate Acquire slabs to locate initial position of canula.
Dynamically re-acquire scan showing position of top of Navigate Alignment canula. With each update show projected target position to determine when alignment is correct.
Insertion Acquire slabs as probe is inserted into brain. Verify that probe is following planned trajectory.
Target Acquire images with probes in place. Review position and redefine target if necessary.
Adjust XY Dynamically re-acquire scan showing position of bottom Offset of canula. With each update show projected target Refine position to determine when offset is correct.
Insertion Acquire slabs as probe is inserted into brain. Verify that probe is following planned trajectory.
Lead Once probe position is finalized, prompt user to place Placement DBS leads or other device.
Admin Admin Re ortin and Archive functionality.
[00871 TABLES 2A-2P provide additional examples of some exemplary operations that may be associated with exemplary workflow steps according to some embodiments of the present invention.
TABLE 2A: Workflow Group Start Step -Desc~i'` tio'n .
The start step can provide UI (User Interface) for selecting procedure laterality.
The start step can provide UI for selecting target: STN, Gpi, or Custom The start step can provide UI for specifying the scanner bore diameter or scanner type that defines the size.
The start step can provide UI for entering hardware identifier data (e.g., version code) from the disposable hardware kit. If the version does not match the version supported by the software, an error can be shown, and the application can remain disabled.
Table 2B: Plan Entry- AC-PC Step Descn tion While data is being sent to the application, the UI can be disabled.
Given a 3D MR series of a whole human head, the application can automatically identify the AC, PC, and MSP points.
The application can display reformatted coronal, sagittal, and axial views aligned to the current AC, PC, and MSP
points.
AC, PC, and MSP points can be editable in any MPR
view in this step. On changing these points, the views can update to realign to the new ACPC coordinate system.
If a new series is sent while in this step, it will replace the existing series and clear all annotations.
While detecting the AC, PC, and MSP points, the UI can be disabled.
If additional data belonging to the current 3D MR series is sent, the AC, PC, and MSP points can be re-calculated automaticall .
[0088] The AC, PC and MSP locations can be identified in any suitable manner. In some embodiments, the AC-PC step can have an automatic, electronic AC, PC MSP Identification Library. The AC, PC and MSP anatomical landmarks define an AC-PC coordinate system, e.g., a Talairach-Tournoux coordination system that can be useful for surgical planning. This library can be used to automatically identify the location of the landmarks. It can be provided as a dynamic linked library that a host application can interface through a set of Application Programming Interface (API) on Microsoft Windows . This library can receive a stack of MR
brain images and fully automatically locates the AC, PC and MSP. The success rate and accuracy can be optimized, and typically it takes a few seconds for the processing. The output is returned as 3D coordinates for AC and PC, and a third point that defines the MSP. This library is purely computation and is typically UI-less. This library can fit a known brain atlas to the MR brain dataset. The utility can be available in form of a dynamic linked library that a host application can interface through a set of Application Programming Interface (API) on Microsoft Windows .
The input to this library can contain the MR brain dataset and can communicate with applications or other servers that include a brain atlas or include a brain atlas (e.g., have an integrated brain atlas). The design can be independent of any particular atlas;
but one suitable atlas is the Cerefy atlas of brain anatomy (note: typically not included in the library). The library can be configured to perform segmentation of the brain and identify certain landmarks. The atlas can then be fitted in 3D to the dataset based on piecewise affine transformation. The output can be a list of vertices of the interested structures.
[0089] In some embodiments, the mid-sagittal plane (MSP) is approximated using several extracted axial slices from the lower part of the input volume, e.g., about 15 equally spaced slices. A brightness equalization can be applied to each slice and an edge mask from each slice can be created using a Canny algorithm. A symmetry axis can be found for each edge mask and identify the actual symmetry axis based on an iterative review and ranking or scoring of tentative symmetry axes. The ranking/scoring cam be based on whether a point on the Canny mask, reflected through the symmetry axis lands on the Canny mask (if so, this axes is scored for that slice). An active appearance model (AAM) can be applied to a brain stem in a reformatted input stack with the defined MSP to identify the AC and PC
points.
[0090] The MSP plane estimate can be refined as well as the AC and PC
points. The MSP plane estimate can be refined using a cropped image with a small region that surrounds a portion of the brain ventricle and an edge mask using a Canny algorithm. The symmetry axis on this edge mask if found following the procedure described above. The AC and PC points are estimated as noted above using the refined MSP and brightness peaks in a small region (e.g., 6x6 mm) around the estimate are searched. The largest peak is the AC point. The PC point can be refined using the PC estimate above and the refined MSP. A Canny edge map of the MSP
image can be computed. Again, a small region (e.g., about 6 mm x 6mm) can be searched for a point that lies on a Canny edge and for which the image gradient is most nearly parallel to the AC-PC direction. The point is moved about 1 mm along the AC-PC direction, towards PC. The largest intensity peak in the direction perpendicular to AC-PC is taken to be the PC point.
Table 2C: Plan Entry - Target Step D;escrr n J
The application can provide the ability to save position coordinates as default values for the STN and Gpi targets. Initially these default values are set to 0,0,0.
Saved values can appear as the default for subsequent procedures using that target.
The application can also provide custom targets for which no default coordinate is supplied. When this option is selected, the user will be able to define a set of custom-named targets associated with a single entry point.
The application can display anatomic coronal, sagittal, and axial views.
Target points can be editable in any MPR view in this step.
The application can provide functionality to automatically register a brain atlas to the patient volume and generate outlines of structures associated with the selected target and display corresponding contours on the MPR views.
The application can provide interface to manually scale and offset the brain atlas registration to better match observed patient anatomy.
Table 2D: Plan Entry - Trajectory Step ., ;
For each given target point, the user can specify the corresponding entry point.
The application can display oblique reformatted coronal, sagittal, and axial views aligned to the proposed trajectory.
Entry points can be editable on either the oblique sagittal or coronal viewports in this step.
The oblique axial view can provide cine functionality to animate a fl -throu h along the trajectory.
The application can also provide an anatomical axial view.
The respective positions of the anatomical and oblique axial views can be represented by lines on the oblique sa ittal and coronal views.
When multiple targets (custom targets) have been defined for an entry point, the application can provide means to select the current target to display. Edits to the entry point will change the entry point for all associated tar ets.
If the user attempts to set the trajectory such that the probe could not be inserted without striking the bore an error can be shown and the trajectory will not be set.
This makes use of the bore size typically entered on the Start step.
The application can provide means to define named trajectories within the step. Trajectories from the list of named trajectories may be selected for display in the step.
If the user moves the entry point off the edge of the grid, warnin text will be shown.
Table 2E: Plan Entry - Grid Step ~'~~esCrl _t10~1 e ?~~~ ~~~~ ?NWW~ ?1 The application can display a volumetric 3D view showing the planning volume. For a bilateral procedure 2 such views can be shown, the left side in the left viewport, the right in the right viewport.
The application can optionally display the grid coordinates of the marking grid.
The application can optionally display overlay graphics to -visually identify edges and positions within grid squares.
On each view, the application can display the corres ondin entry point The application can automatically align the 3D views such that the user's point of view is looking along the trajectory from the entry towards the target. In the case of multiple custom targets, the trajectory to the first target can be used.
The application can set the visualization parameters of the volume such that the grid itself is visible to the user.
The initial zoom level can ensure that the entire head is visible.
100911 At this point, holes have been burred at the entry points and the trajectory guides 50t have been attached. NOTE: Because the patient has been moved, points defined in the previous image coordinate system may no longer be valid.
Also, brain shift may occur at this point.
Table 2F: Plan Target - AC-PC Step Descri pti The Plan Target AC-PC step can look and function the same as in the planning AC-PC step. However, data received in this step can be stored as the replanning volume. The AC, PC, and MSP annotations and the resultant transformation derived from the replanning volume can be kept distinct from those determined in the planning AC-PC step.
Table 2G: Plan Target- Target Step , ,. .
The Plan Target Step can function the same as in the planning step but with additional functionality to support slab data fusion.
The Plan Target Step can accept DICOM slab data.
While receiving slab data, the UI can be disabled and a message should be shown to indicate that a data transfer is in ro ress.
The Plan Target Step can provide a thumbnail bar that lists series in their order of acquisition. Selecting a series in the thumbnail bar will cause it to appear fused with the re-planning volume. Selecting the Plan Target volume can cause it to be dis la ed by itself.
Fused data can appear in the viewports along with the -plan target volume images and will be positioned and scaled to exactly coincide with the position and scale of the plan target volume.
The control panel can contain a slider that controls the relative intensity of the two series in the blended viewports.
The step can display scan plane parameters for an anatomical axial slab through the current target.
Table 2H: Plan Target - Trajectory Review Step Re,, uireme.nt des'cn tion The Plan Target Trajectory Review Step can function the same as the Planning Trajectory review step with only the following exceptions:
- Slab fusion support - Segment out the pivot point from frame markers - Use pivot point position as a fixed entry point (not editable) The Plan Target Trajectory Step can accept DICOM slab data.
While receiving slab data, the Ul can be disabled.
The Plan Target Trajectory Review Step can provide a thumbnail bar that lists series in their order of acquisition.
Selecting a series in the thumbnail bar will cause it to appear fused with the re-planning volume. Selecting the re-planning volume can cause it to be dis la ed by itself.
Fused data will appear in the viewports along with the volume images and will be positioned and scaled to exactly coincide with the position and scale of the re-planning volume.
The control panel can contain a slider that controls the relative intensity of each series in the blended viewports.
If for any reason the software is unable to identify the frame markers to find the pivot point, a warning can be dis ia ed.
The step can display scan plane parameters for:
- an oblique sagittal slab along the trajectory - an oblique coronal slab along the trajectory The step can display the trajectory angles relative to the anatomical coronal and sagittal planes.
Table 21: Navigate - Initiate ;Descr~ t_~on .
The step can prompt the user to acquire a small high-resolution slab through the proximal canula at a distance such that it will show a cross-section of the proximal canula even at maximum angulation and maximum offset. The slab can have a minimum of 4 slices.
(Example: given a canula 83 mm long, a maximum angulation of +/- 35 degrees, and a maximum offset of 4mm, then a scan plane 65 mm up from the distal canula marker will be sufficient to ensure that the canula is visible in the slab.) The application can also prompt the user to acquire a small high resolution slab scan with the following attributes:
= plane aligned to the plane of the frame markers (this can be based on the frame segmentation that was done for the trajectory review step) = plane center is positioned at the mechanical center of rotation = slab thickness is large enough to include all of the distal canula marker even under maximum angulation and offset (Example: given a maximum angulation +1-35 degrees, a maximum offset of 4 mm, and distal canula marker size of 7.1 mm, the total thickness required would be 11.6 mm, so any larger value may be used, say 13 mm) = slice spacing can be about 1 mm The application can identify the positions of the proximal and distal canula.
Using the detected positions of the frame markers in the plan target volume, the application can compare the observed position of the distal canula with the mechanical center of rotation. Since a locking pin may be used to ensure that there is no offset, values above a low threshold can cause a warning to be displayed.
(Example warning text: "Distal canula marker not found at expected location. Verify that canula is locked in `down' osition and reacquire distal canula scan."
If no pivot point marker can be identified, the user will be prompted to verify that the canula is locked in the 'down' position and re-acquire the slab scan.
The step can provide 3 MPR viewports in which to display the acquired slabs. These viewports will be oriented such that their base planes are aligned to the detected canula axis.
The step can provide a thumbnail bar to allow the user to select which ac uired slab to dis la .
Table 2J: Navigate - Alignment Descr,-i = tion The application can prompt for an alignment scan with the following attributes:
= scan plane is perpendicular to planned trajectory = scan plane is centered around the trajectory = scan plane position is set such that a cross-section of the proximal canula will be shown even at maximum angulation and maximum offset.
= a single 2D image can be acquired The application can display an anatomical axial view through the currently selected target.
The user may opt to switch the display to show a tra'ecto -axial view.
On receiving a 2D image of the top of the canula, the application can automatically identify the position of the top of the canula in 3D space. Using this position and, previously-determined pivot point, and the previously-determine offset, the application can draw an annotation representing the intersection of the current trajectory with the image plane containing the plahned target.
This step can display lines from the current projected target to the planned target that indicate the track the projected target would travel if the pitch and roll wheels were turned independently. These lines can be colored to match colors on the control wheels for pitch and roll respectively. A tool-tip (e.g., pop-up) can provide text to describe the suggested action.
For example: "Turn Roll knob to the Left"
On re-calculating the projected target point, an error value can indicate the in-plane linear distance between the projected target point and the planned target point on the currentl -dis la ed plane.
Images that are not oriented correctly to the requested scan plane can result in a warning. In this case, annotations and error measurements may not be dis la ed.
Images in which the targeting canula cannot be identified can result in a warning. Annotations and error measurements may not be dis la ed.
When multiple targets have been defined for an entry point, the application can provide means to select the current trajectory to dis la .
When drawing the target and the current projection of the canula path, the annotations can be drawn to match the physical size of the probe diameter.
Table 2K: Navigate - Insertion Desc:n tion The application can provide a depth value to set on the depth stop prior to insertion.
The application can prompt with scan parameters for oblique coronal and sagittal planes aligned to the trajectory. Also for an oblique axial perpendicular to the tra'ecto .
On receiving coronal or sagittal images, the application can display an overlay graphic indicating the planned trajectory. The most recent coronal and sagittal images can appear together in a 1 x2 dis la .
On receiving a trajectory axial scan perpendicular to the trajectory, the application can segment out the cross-sections of the probe to determine the actual path being followed by the probe.
On receiving a trajectory axial scan perpendicular to the trajectory, the application can display two viewports containing:
= the axial stack with graphic overlays showing the detected path of the probe on each image = an anatomic axial view through the target showing the planned target and the target projected from the detected path of the probe. An error value can show the distance between the current projected target and the planned target.
If multiple trajectories have been defined for a single entry, the application can display the trajectory that is currently aligned during insertion.
On entering the insertion step, the application can instruct the user to ensure that if they are using an imaging probe that it is connected to the scanner as an internal coil. Failure to do so could cause heating of the coil and injury to the patient. The user must explicitly click a button to acknowledge that they understand the warning.
Table 2L: Refine - Target The step can prompt for either i) a high-resolution 2D
image to be acquired using the imaging probe or ii) a high-resolution slab through the target area. The associated scan plane parameters can specify a trajectory axial image centered on the target.
The application can provide means to identify the tip of the probe (or stylet), and provide an error value for the linear distance from the probe to the planned target point in the axial anatomic plane.
The application can provide UI to set an updated target point.
The user may opt to proceed to the X-Y Adjustment step, return to Alignment to align to another target, or advance to the Admin step.
On accepting the current position, the user can be shown a warning not to scan once the MR-incompatible DBS
leads (if they are incompatible or potentially incompatible) have been placed. For bilateral cases, the user proceeds to complete the insertion of the probe on both sides before placing the leads. The message can indicate that scanning with MR-incompatible leads may result in serious injury or death.
If the user modifies a target point, the step can prompt the user to confirm removal of the offset locking pin. from the targeting frame before going on to the next step.
[0092] In the event that the placement is not acceptable, the user may opt to proceed to the X-Y Adjustment workflow step.
Table 2M: Refine - Adjust X-Y Offset ~ D~esc,yi tion The X-Y Adjustment step can display the current target and projected point as annotations to the image data that was acquired during the Target Refinement step.
This step can prompt the user to acquire 2D images with scan plane parameters such that the image lies perpendicular to the trajectory and through the pivot oint.
On receiving a 2D image through the pivot point, the step can calculate the current projected target and display an annotation on the 2D image from the imaging probe.
This step can display lines from the current projected target to the revised target that indicate the track the projected target would travel if the X and Y offset wheels were turned independently. The lines can be colored to match colors on the control wheels for X and Y offset respectively. A tool-tip (e.g., pop-up) can provide text to describe the necessary action.
(For example: "Turn X-offset knob to the Left"
This step can display an annotation indicating the location of the original planned target.
When drawing the target and the current projection of the canula path, the annotations can be drawn to match the physical size of the probe diameter.
Table 2N: Refine - Insertion Desc,r~ .tion .
After completing the X-Y Adjustment, the application can provide a workflow step to guide insertion. This is substantially the same as the first instance of the Insertion step above.
Table 20: Refine - Lead Placement , ;~ . .. _., D;esc After probe has been placed (or both probes for bilateral case) and position has been accepted by user, the user may proceed to the lead placementstep.
This step can provide the user with the depth values for the placement of the leads.
This step may advise the user that once leads have been placed scans may not be performed because heating in the leads could cause injury or death to the patient.
Table 2P: Admin Step Descn:t~on The step can provide means to archive data relating to the procedure. This includes:
= trajectory planning data = log= files with case data = image data The step can provide functionality to automatically generate a report documenting the performed procedure.
This report can include:
= patient information = AC, PC, and MSP points in MR space (Both detected and user-specified, if user modified.) = Planned and corrected targets in both MR and ACPC space = elapsed time for the procedure = physician case notes (optional) = any screenshots taken during the procedure An anonymous version of the report can also be generated automatically with the patient name and id removed.
The step can provide UI whereby the user can:
= selected a target = specify a position in MR coordinates representing the lead tip = define a set of offsets indicating the electrode offsets from the lead tip For each offset value, the step can find provide the ACPC coordinate that corresponds to a point offset from the tip position back along the trajectory of the lead.
These values correspond to the electrode positions in ACPC space. These values may be added to a patient report.
The Admin step can include a button to shut down the application on completion of the procedure. The user may not be allowed to otherwise close the application.
The application can have a configuration value to specify whether all patient data is to be cleared from the system on shut down.
Again, it is noted that functional patient data can be obtained in real-time and provided to the circuit 30c/workstation 30 on the display 32 with the visualizations of the patient anatomy to help in refining or planning a trajectory and/or target location for a surgical procedure.
[0093] When displaying images or visualizations that were created with the imaging probe 50a (Figure 5, where used), the circuit 30c may electronically apply a (sigma) correction to correct for a`volcano' or 'halo' shaped intensity distortion. That is, in some particular embodiments, the imaging probe antenna or coil may introduce a distortion in the images that are created with it that may be described as a bright halo around the probe itself where it appears in the image. Thus, when images from the imaging probe have such a feature, the circuit 30c can be configured to electronically automatically apply a correction to cancel out the halo (or at least reduce it). This feature will only affect the small field, high-resolution images that are created via the imaging probe itself. Images obtained using the main head coil do not typically have such a distortion.
10094] Referring to Figure 14A, the circuit 30c can be configured with a single control tool bar 30t that is displayed on the display 32 that allows the user to select what group and step to go to and also shows which groups and steps have been completed. Figure 14B shows an example of the workflow control tool bar 30t with the "Plan Entry" group selected, and the "Define Target" enlarged as the current step.
The tool bar 30t can include a color border 130 that can be used to partially or totally surround a button 130b to illustrate completion of a step. For bilateral procedures, the border can be color enhanced on one side when a task for that side is completed, e.g., on the left side when the left target is complete in the Define Target step.
[0095] As the user works through the procedure, certain clinical information is stored to be incorporated into a procedure report that may be reviewed at the end of the procedure and/or archived as a patient record. The circuit 30c can be configured to provide a user interface (UI) 301 that provides viewing tools, such as one or more of the following features.
- Draw measurement lines - Pan view - Zoom, Zoom All, Zoom to Region, Zoom to Point - Magnifying glass - Show/Hide Annotations - Show/Hide Crosshairs - Drag views between panels - Resize view panels - Maximize view to a lxl display - Save screen capture (can be added to the report) - Reset view settings to default [0096] Figure 15 is a screen shot of an exemplary UI 301 for the Start Group which may conveniently be configured as a one-screen input to set overall procedure parameters such as laterality, target type and MR Scanner bore size (recognizing that open bore MRI Scanner systems may also be used). Instead of bore size, a drop-down list can be provided that allows a user to select a manufacturer and type of MR Scanner in use that provides the associated bore size. Of course, the system 10 can be configured as an MR Scanner-specific system or the MR
Scanners of the future may have a standard bore size or be configured so that bore size is not a constraint and this information may not be required. As shown, the UI 301 can also include an input 321 that requires entry of hardware identifier data 66, shown as disposable kit version input, as discussed above with respect to Figure 5. In order to assure hardware and software compatibility and/or proper operation, if the identifier data 66 does not match, the system 10 can be configured to not allow a user to proceed to a next step or may prompt the user for other key codes.
[0097] Bore size is used in the step Plan Entry / Trajectory. If the user selects a trajectory such that the probe cannot be inserted into the canula 60 because it will not physically fit inside the scanner bore, a warning is generated (visual on the display 32 and/or audible). See Plan Entry / Trajectory step above.
[0098] Figure 16 is a screen shot associated with an exemplary UI 301 with a Plan Entry workflow group for a Define Target step (shown as left STN) illustrating the toolbar 30t and cross hair location data with default landmarks with patient image data. This step can be used to establish the AC PC coordinate system.
On receiving a whole-head volume, the step can automatically find candidates for these points. The user is to review and correct these points if necessary. The user can either position the crosshairs at a point and "click" the `Set' button to set=
the desired annotation, or they can drag an existing annotation around on the screen. Once points are defined, view planes can automatically reformat to align them to the ACPC
coordinate system to show the anatomical planes: Coronal, Sagittal, and Axial.
Any subsequent edit to the landmarks can cause the view planes to instantly re-align to match.
[0099J Figure 17 illustrates another exemplary screen shot for a Define Target step which is used to set target points so that the trajectories through potential entry points can be investigated in the next step. The user may opt to overlay the outlines from a standard brain atlas over the patient anatomy for comparison purposes which may b provided in color with different colors for different structures.
Figure 17 shows the play entry of the Define Target step with no atlas. Figure 18 illustrates the UI with an atlas showing a target outline in three orthogonal views and Figure 19 illustrates an atlas showing structures in a 1 x 1 layout. The view planes show the anatomical planes as defined in the ACPC step. Target points are edited similarly to how the ACPC points are edited. For the bilateral case, once the target has been defined for one side, then when the user selects the target for the other side, the crosshairs will automatically jump to the mirror-image position. If the patient has symmetric anatomy, this will save time in finding the equivalent position.
When using the brain atlas, the user may opt to show either just the target structure (STN or GPi) or all structures. In either case, a tooltip (e.g., pop-up) can help the user to identify unfamiliar structures. The user may also opt to scale and/or shift the brain atlas relative to the patient image to make a better match. To do this, the user may drag the white outline surrounding the brain atlas template. Fiber track structures and/or functional information of a patient's brain can be provided in a visually prominent manner (e.g., color coded or other visual presentation) for a surgeon's ease of reference. This information can also be selected or suppressed from views via a UI
selection 32F (e.g., toolbar option) as shown in Figure 4.
[01001 Figure 20 is a screen shot of an exemplary screen display for a Trajectory step in the Plan Entry workflow group. This step is used to find a clinically viable trajectory that determines the entry point on the skull typically via a grid such as that provided by the grid patch 50g (Figure 11). A visual warning can be displayed, e.g., a red warning message 30W on the top of the two top views and a red trajectory line 30R) can be used to indicate that the selected trajectory does not intersect the grid. In operation, upon entering this step, the workstation 30 can automatically search through the image volume for the marking grid(s) 50g. It can be configured to position the initial trajectory such that it runs through the middle of the grid 50g. If the user moves the trajectory such that the entry point is not on the grid, a warning 30W is displayed.
[0101] The top two views of Figure 20 show the coronal and sagittal views aligned along the current trajectory line. The user drags the trajectory line and it rotates freely about the target point. The bottom-left view shows the plane perpendicular to the trajectory at the level of the green line in the coronal and sagittal views. This is the "probe's eye" view. To advance the probe's eye view along the trajectory, in addition to dragging the green line along the trajectory line or using the mousewheel, onscreen VCR-style controls can provide an animated fly-through.
[0102] Figure 21 illustrates a (pop-up) warning 30W' that is automatically generated when a user selects a trajectory that may be blocked by the scanner bore wall. That is, if the user sets a trajectory such that the scanner bore will interfere with the insertion of the probe, a warning is displayed. This calculation is based on the bore size, current trajectory angles, and pre-configured values for the size of the probe 50a/50s/52, canula 60, and frame/trajectory guide 50t (Figure 5). Similar to the target step, in the bilateral case once the trajectory has been set on one side, the initial default for the other side can be a mirror-image trajectory to start the user closer to a more a likely trajectory. On saving the trajectory, the step automatically finds the surface of the skull along the planned/selected trajectory, identifies the coordinates on the grid and stores that location as the entry point.
[0103] Figure 22 illustrates an exemplary grid 50g shown overlying a patient's skull on a display 32 (without annotation lines), illustrating coordinates for selecting an entry location (shown as columns 1-6 and rows A-F) with a left STN
entry location. Figure 23 shows the same screen view as Figure 22 but with optional annotation grid lines. Typically, a user can see the grid coordinates clearly enough that the optional overlay grid lines are not required in order to identify the grid elements. However, in some embodiments, as shown a UI will allow a user to display the lines.
[0104] Figure 24 is an exemplary screen shot of a Plan Target Group workflow with the ACPC step selected. This workflow group is used to precisely determine target points after one burr hole has been formed for a unilateral procedure or both burr-holes have been formed for bilateral procedures (the surgical entries have been burred and the frames attached). Previous planning is typically invalidated by brain shift that occurs with the loss of cerebral spinal fluid pressure. This step is substantially similar to the Plan Entry ACPC step discussed above. The difference is that in addition to the whole-head volume, the user may optionally also send one or more high-resolution slabs containing the ACPC points. The user can use a thumbnail bar to select which data to use to display and edit the ACPC landmarks. Figure illustrates a screen shot with a Plan Target/ACPC step showing slab data.
[0105] Figure 26 is an exemplary screen shot of a Plan Target Group workflow with the Define Target step selected. The brain images can be shown with blended volume and slab image data. This step is similar to the Define Target step in the Plan Entry group, but like the preceding step it also supports high-resolution slabs to increase the precision with which the target can be defined. Thus, this step has the ability to show a blended image using data from both the volume and a high-resolution slab. A slider 30S (shown to right under the word "blend") in the UI allows the user to control the ratio of how much of the image comes from either source.
[0106] Figure 27 is a screen shot of a Trajectory step in the Plan Target workflow group. This step is similar to the Plan Entry/Trajectory step, but in this case this step is typically only used to review the trajectory. The entry point cannot be changed since the hole has been burred and the frame attached. Also, like the preceding step, this can support the use of high-resolution slabs and can blend between the volume and the slab. On entering this step, the software can automatically search through the whole-head volume and find the frame marker fiducials. From these, the system 10 and/or circuit 30c can determine the frame locations and orientation and uses this to calculate the actual entry point on the head.
This is used along with the target from the previous step to determine the trajectory.
Otherwise, the trajectory may be reviewed as described above in the Plan Entry /
Trajectory step.
[0107] Figure 28 is a screen shot of a Navigate workflow group with the Initiate Navigation step selected. This visualization shows the proximal canula slab.
On entering this group, the user has already burred the entry hole, attached the frame, and finalized their planned targets. This group will guide the user through aligning the canula 60 (Figure 8) to match the planned trajectory and inserting the probe through the canula to the correct depth. This step is used to determine the initial physical location of the canula. The user acquires a scan through the top of the canula and another through the bottom. The circuit 30c (e.g., software) automatically finds the canula in each slab and determines its position and orientation in space.
Figure 29 illustrates the distal canula slab in the Navigate/Initiate Navigation step.
[0108] Figure 30 is a screen shot of a Navigate workflow group with the Align Canula step selected. This step is used to physically align the canula .5 (Figure 8) to the planned trajectory. The user iteratively adjusts the canula angulation via the pitch and roll control knobs as they rapidly re-acquire an image through the top of the canula 60p. With each update, the circuit 30c (e.g., software) calculates the position of the canula 60 and displays an annotation showing where it's currently pointing on the target plane. Figure 30 illustrates that user feedback such as a prompt over overlay or a tooltip-style pop-up can tell the user which control to turn and which direction. In some embodiments, as shown, annotations can be drawn as circles corresponding to the probe diameter. The circle with the crosshair is the planned target, the other circle is the current projected point based on the trajectory of the canula. The lines in between show the relative amount of pitch and roll to apply and the text can specify which known to turn and in which direction (shown as "turn pitch wheel to the left").
[0109] Figure 31 illustrates a screen shot from an exemplary Navigate/Insert Probe step. This screen shot illustrates coronal and sagittal views to the target (e.g., STN) and can provide a set depth stop dimension (shown on the upper right hand side of the UI). This step allows the user to see how well the probe is following the trajectory as it is inserted. The user may opt to scan Coronal and Sagittal slabs along the probe to visually determine the probe alignment in those planes. The user can also scan perpendicular to the probe. In that case, the circuit 30c (e.g., software) can automatically identify where the probe is in the slab and it then shows a projection of the current path onto the target plane to indicate the degree and direction of error if the current path is continued. Figure 32 illustrates an axial slab and projected point with a projected error if the current trajectory continues (the probe is shown in the right image offset by 4.6 mm and the offset is also noted on the UI as "Projected Error". The user can perform these scans multiple times during the insertion. The automatic segmentation of the probe and the display of the projected target on the target plane provide fully-automatic support for verifying the current path. The Coronal/Sagittal views can provide the physician with a visual confirmation of the probe path that doesn't depend on software segmentation.
[01101 Figure 33 illustrates an exemplary screen shot of a Refine Placement workflow group with the Target Revision step selected. This step can illustrate the target slab. After completing the initial insertion, the user (e.g., physician) may find that either the placement doesn't correspond sufficiently close or perfectly to the plan, or the plan was not correct. This may be particularly likely if an imaging probe (50a) is used, since a user will be able to more clearly visualize structures like the STN that are usually indistinct with external coils. This workflow group can support functionality whereby the physician can withdraw the probe and use the X and Y offset adjustments to obtain a parallel trajectory to a revised target.
This step can prompt the user or otherwise acquire an image slab through the distal tip of the probe. (Optionally, this step may use the imaging probe). The step displays the slab and on it the user may opt to modify the target point to a new location or accept the current position as final.
[0111] Figure 34 illustrates an exemplary screen shot of the Refine Placement workflow group with the Adjust X-Y Offset step selected. This step is very similar to the Navigate/Align Canula step described above. The primary difference is that instead of adjusting the angulation of the canula 60, the user is adjusting a small X-Y offset to set the canula 60 to a trajectory parallel to the original one. Figure 34 shows the display with an visualization f the position of the probe tip relative to the target and with instructions on what physical adjustments to make to obtain the desired parallel trajectory (shown as "turn Y wheel to the right") and the projected error (shown a 2.7 mm). Figure 35 illustrates a detail of the adjust annotations and pop-up (shown as "turn X wheel to the left").
[0112] After the X-Y adjustments are made, the Insert Probe of the Refine Placement workflow group is selected and carried out in the same manner as the Navigate/Insert Probe step described above.
[0113] Figure 36 illustrates the Refine Placement workflow group with the "conclude procedure" step selected. This step occurs after all probes have'been inserted and have had their positions verified by the physician. At this point, the U1 can prompt them to insert both leads where implantable stimulation leads are to be placed (using the defined trajectory) and can warn them not to perform any additional scans if MRI-incompatible or potentially incompatible leads are used. As shown, the system 10 can be configured to define (and output to a user) or depth stops to set the lead or other therapy or diagnostic device for each STN or target site. The depth stops can be different for each implant location on the left and right targets (for bilateral procedures) so that the electrodes of the leads or other components for other devices are positioned in the desired location.
[0114] Figure 37 illustrates an example of a display that may be used for the ADMIN workflow group. This group has one step that provides reporting and archive functionality. The report automatically documents the entire procedure including annotations, measurements, and screen captures. The circuit 30c can generate a full version and an anonymous version of the report and may include a date as to when everything is archived to CD.
[0115] The circuit 30c may also be configured to determine where individual electrodes on the DBS leads are situated in ACPC coordinates. Given the tip position in MR coordinates (the circuit 30c can ill in the planned position, but the user may change it) the user will provide a set of offset values that represent the distance of each electrode from the lead tip. In other embodiments, a lead type can be selected such as from a pull-down list and those values can automatically be input based on the manufacturer and lead type (e.g., design thereof). The circuit 30c can be configured so that the UI displays the corresponding electrode positions in ACPC
coordinates.
[0116] Figure 38 illustrates an example of a display with a UI that may be used for the ADMIN workflow group shown as- Admin Page/Electrode Offset dialog step selected. This step may also be included in the Conclude Procedure/step or provided as a separate workflow group. The electrode offset values may significantly speed up the process by which the pulse generator is programmed since the physician will know where the electrodes are anatomically.
[0117] The system 10 may also include a decoupling/tuning circuit that allows the system to cooperate with an MRI scanner 20 and filters and the like. See, e.g., U.S. Patent Nos. 6,701,176; 6, 904,307 and U.S. Patent Application Publication No. 2003/0050557, the contents of which are hereby incorporated by reference as if recited in full herein. As noted above, one or more of the tools can include an intrabody MRI antenna 50a (Figure 5) that is configured to pick-up MRI signals in local tissue during an MRI procedure. The MRI antenna can be configured to reside on the distal portion of the probe. In some embodiments, the antenna has a focal length or signal-receiving length of between about 1-5 cm, and typically is configured to have a viewing length to receive MRI signals from local tissue of between about 1-2.5 cm. The MRI antenna can be formed as comprising a coaxial and/or triaxial antenna. However, other antenna configurations can be used, such as, for example, a whip antenna, a coil antenna, a loopless antenna, and/or a looped antenna.
See, e.g., U.S. Patent Nos. 5,699,801; 5,928,145; 6,263,229; 6,606,513; 6,628,980;
6,284,971;
6,675,033; and 6,701,176, the contents of which are hereby incorporated by reference as if recited in full herein. See also U.S. Patent Application Publication Nos.
2003/0050557; 2004/0046557; and 2003/0028095, the contents of which are also hereby incorporated by reference as if recited in full herein.
(0118] In some embodiments, the implanted leads and/or intrabody tools can be configured to allow for safe MRI operation so as to reduce the likelihood of undesired deposition of current or voltage in tissue. The leads or tools can include RF
chokes such as a series of axially spaced apart Balun circuits or other suitable circuit configurations. See, e.g., U.S. Patent No. 6,284,971, the contents of which are hereby incorporated by reference as if recited in full herein, for additional description of RF
inhibiting coaxial cable that can inhibit RF induced current. The conductors connecting electrodes or other components on or in the tools can also include a series of back and forth segments (e.g., the lead can turn on itself in a lengthwise direction a number of times along its length) and/or include high impedance circuits. See, e.g., U.S. Patent Application Serial Nos. 11/417,594; 12/047,602; and 12/090,583, the contents of which are hereby incorporated by reference as if recited in full herein.
10119] Although not shown, in some embodiments, one or more of the surgical tools can be configured with one or more lumens and exit ports that deliver desired cellular, biological, and/or drug therapeutics to the target area, such as the brain. The tools may also incorporate transseptal needles, biopsy and/or injection needles as well as ablation means. The lumens, where used, may receive extendable needles that may exit the probe from the distal end or from the sides, proximal, distal, or even, through the electrodes to precisely deliver cellular/biological therapeutics to the desired anatomy target. This delivery configuration may be a potential way to treat patients, where the cellular/biological therapeutics can be delivered into the desired anatomy to modify their cellular function. The cells (e.g., stem cells) may improve function. MRI can typically be effectively used to monitor the efficacy and/or delivery of the therapy.
[01201 The system 10 can include circuits and/modules that can comprise computer program code used to automatically or semi-automatically carry out operations to generate visualizations and provide output to a user to facilitate MRI-guided diagnostic and therapy procedures. Figure 39 is a schematic illustration of a circuit or data processing system that can be used with the system 10. The circuits and/or data processing systems may be incorporated in one or more digital signal processors in any suitable device or devices. As shown in Figure 39, the processor 410 communicates with an MRI scanner 20 and with memory 414 via an address/data bus 448. The processor 410 can be any commercially available or custom microprocessor. The memory 414 is representative of the overall hierarchy of memory devices containing the software and data used to implement the functionality of the data processing system. The memory 414 can include, but is not limited to, the following types of devices: cache, ROM, PROM, EPROM, EEPROM, flash memory, SRAM, and DRAM.
[0121] As shown in Figure 39 illustrates that the memory 414 may include several categories of software and data used in the data processing system:
the operating system 452; the application programs 454; the input/output (I/O) device drivers 458; and data 456. The data 456 can also include predefined characteristics of different surgical tools and patient image data 455. Figure 39 also illustrates the application programs 454 can include a Visualization Module 450, Interventional Tool Data Module 451, a Tool Segmentation Module 452 (such as segmentation modules for a grid patch, a targeting canula, and a trajectory guide frame and/or base), and a workflow group User Interface Module 453 (that facilitates user actions and provides guidance to obtain a desired trajectory such as physical adjustments to achieve same).
[01221 As will be appreciated by those of skill in the art, the operating systems 452 may be any operating system suitable for use with a data processing system, such as OS/2, AIX, DOS, OS/390 or System390 from International Business Machines Corporation, Armonk, NY, Windows CE, Windows NT, Windows95, Windows98, Windows2000 or other Windows versions from Microsoft Corporation, Redmond, WA, Unix or Linux or FreeBSD, Palm OS from Palm, Inc., Mac OS from Apple Computer, LabView, or proprietary operating systems. The I/O device drivers 458 typically include software routines accessed through the operating system 452 by the application programs 454 to communicate with devices such as I/0 data port(s), data storage 456 and certain memory 414 components. The application programs are illustrative of the programs that implement the various features of the data processing system and can include at least one application, which supports operations according to embodiments of the present invention. Finally, the data 456 represents the static and dynamic data used by the application programs 454, the operating system 452, the I/O device drivers 458, and other software programs that may reside in the memory 414.
[0123] While the present invention is illustrated, for example, with reference to the Modules 450-453 being application programs in Figure 39, as will be appreciated by those of skill in the art, other configurations may also be utilized while still benefiting from the teachings of the present invention. For example, the Modules 450-453 and/or may also be incorporated into the operating system 452, the I/O
device drivers 458 or other such logical division of the data processing system. Thus, the present invention should not be construed as limited to the configuration of Figure 39 which is intended to encompass any configuration capable of carrying out the operations described herein. Further, one or more of modules, i.e., Modules 450-453 can communicate with or be incorporated totally or partially in other components, such as a workstation, an MRI scanner, an interface device. Typically, the workstation 30 will include the modules 450-453 and the MR scanner with include a module that communicates wit the workstation 30 and can push image data thereto.
[0124] The I/O data port can be used to transfer information between the data processing system, the circuit 30c or workstation 30, the MRI scanner 20, and another computer system or a network (e.g., the Internet) or to other devices controlled by or in communication with the processor. These components may be conventional components such as those used in many conventional data processing systems, which may be configured in accordance with the present invention to operate as described herein.
[0125] In the drawings and specification, there have been disclosed embodiments of the invention and, although specific terms are employed, they are used in a generic and descriptive sense only and not for purposes of limitation, the scope of the invention being set forth in the following claims. Thus, the foregoing is illustrative of the present invention and is not to be construed as limiting thereof.
More particularly, the workflow steps may be carried out in a different manner, in a different order and/or with other workflow steps or may omit some or replace some workflow steps with other steps. Although a few exemplary embodiments of this invention have been described, those skilled in the art will readily appreciate that many modifications are possible in the exemplary embodiments without materially departing from the novel teachings and advantages of this invention.
Accordingly, all such modifications are intended to be included within the scope of this invention as defined in the claims. In the claims, means-plus-function clauses, where used, are intended to cover the structures described herein as performing the recited function and not only structural equivalents but also equivalent structures. Therefore, it is to be understood that the foregoing is illustrative of the present invention and is not to be construed as limited to the specific embodiments disclosed, and that modifications to the disclosed embodiments, as well as other embodiments, are intended to be included within the scope of the appended claims. The invention is defined by the following claims, with equivalents of the claims to be included therein.
For additional discussion of suitable trajectory guides, see, U.S. Application Serial No.
12/134,412, and co-pending, co-assigned U.S. Patent Application Serial No.
identified by Attorney Docket No. 9450-341P, the contents of which are hereby incorporated by reference as if recited in full herein.
[0075] Figures 6B and 6C illustrate examples of a depth stop 210 cooperating with an elongate member 212 such as, for example, a stimulation or ablation lead, a diagnostic and/or imaging probe, a removable sheath and/or other therapy or diagnostic device inserted and secured therein as illustrated. The illustrated depth stop 210 has a generally cylindrical configuration with opposite proximal and distal ends 210a, 210b and is adapted to be removably secured within the proximal end of the tubular trajectory guide member 50t (Figure 6A). The depth stop 210 is configured to limit a distance that the member 212 extends into the body of a patient when the depth stop is inserted within the tubular member 50t or 60. The member 212 can include visual indicia of insertion depth 215 to allow a user to visually attach the stop 210 at the appropriate location that provides the desired insertion depth.
[0076] As shown in Figures 6C and 6D, in some embodiments, the depth stop 210 is attached to a peel-away sheath 212s (see also probe 50s, Figure 5) and can be configured to receive and guide an elongated interventional device, such as a stylet or imaging probe, therethrough. As shown in Figures 6C and 6D, the sheath 212s can include opposing tabs that, when pulled apart, cause the sheath to peel away for removal from the targeting canula 60. In other embodiments, the depth stop 210 can be attached to a stimulation lead to allow for a defined insertion depth (see, e.g., Figure 38 below) or other device where insertion depth control is desired.
[0077] As shown in Figure 7, the targeting canula 60 can be attached to the trajectory guide 50t. The targeting canula 60 can include a through passage 61.
The distal end of the targeting canula 60 has a fiducial marker 60m, shown as a substantially spherical or round (cross-section) marker shape. The proximal end 60p can be configured with a fluid filled channel 68 concentric with the passage 61 that can define a cylindrical fiducial marker. Along the axis of the canula 60, there is the lumen or passage 61 through which another device can be slidably introduced and/or withdrawn, e.g., a stylet 50s, imaging probe 50a, therapy delivery or diagnostic device 52 (Figure 5) and DBS stimulation leads for implantation, can be advanced into the brain.
[0078] Figure 11A illustrates an example of a grid patch 50g with physical characteristics that are predefined and available to or in the circuit 30c (e.g., software application). Figure 11B illustrates that the grid can allow for precise correlation of logical points in an MR volume with the physical patient. The system 10 can be configured to generate a 3-D volumetric view on the display 32 with overlays to show the grid 50g and an entry point and grid coordinates (and optionally grid edges) allowing a burr hole mark to be made on a skull at the desired planned entry point.
[0079] Figures 12A-12E illustrate a series of operations that can be carried out for the grid 50g segmentation. Some or all of these operations may be carried out behind the scene (e.g., not actually displayed). Figure 12A
illustrates an example of a histogram. Figure 12B is an example of an initial distance image.
Figure 12C is an image of a result of searching for the grid in the initial distance image. Figure 12D illustrates the result of searching for the grid in the optimal distance image, with small step sizes. Figure 12E illustrates the result of spatially deforming the grid to fit the head surface and interpolating the grid cells to find the vertices.
[0080] With reference to Figure 12A, the amplitude of the background noise in the input image stack can be estimated. To do so, a histogram of the stack can be constructed. The first negative maximum of the slope of the histogram can be located. The first peak to the left of this maximum can be located (this can be termed the "noise peak"). The difference between the noise peak and the first negative maximum of the slope can approximate the standard deviation of the noise. The noise threshold can be obtained using the following Equation (1).
(noise threshold) = (noise peak) + 4*(noise standard deviation) EQUATION (1) [0081] The above can be considered as a first step in the grid segmentation of the image data. Steps 2-5 can be carried out as described below to place the grid in position and deform to curvature of the skull for the grid segmentation.
2. Estimate the "optimal view direction" for the grid. This is defined as the vector from the grid center to the midpoint of the AC-PC line:
a. Construct an initial distance image as shown in Figure 12B starting from a rough estimate of the optimal view direction:
i. Create an image plane that is perpendicular to the estimated view direction, and outside of the head volume.
ii. For each pixel in the image plane, record its distance to the first voxel in the image volume that is above the noise threshold.
b. Find the grid cells in the distance image:
i. The grid cells are "bumps" on the head surface, so they appear as local minima in the distance image. This characteristic is enhanced by applying a Laplacian-of-Gaussian operator to the distance image.
ii. For each location in the image, place a virtual grid (whose dimensions match those of the physical grid), at several orientations.
Compute a score by tallying 1 point for each virtual grid cell that lands on one of the image minima. The highest score corresponds to the location and orientation of the grid in the distance image.
c. The center of the grid so obtained gives a new estimate of the optimal view direction. This estimate is used as a starting point of a new iteration of steps 2a, and 2b, in order to refine the estimate.
3. Find the grid using the optimal view direction:
a. Construct a distance-image as described in step 2a, but using the optimal view direction obtained in step 2.
b. Find the grid in the new distance image, as described in 2b.
c. Refine the grid position and orientation by repeating the search procedure of 2b, but search only in a small region around the known approximate location, with very fine step-sizes (see, Figure 12D).
4. Deform the grid to the shape of the head:
a. Up to this point, the grid has been taken to be planar.
b. Fit the grid cell locations in the distance image to a cubic surface using a robust regression algorithm, and snap each grid cell from step 3 to this surface. (This is to avoid producing a very irregular grid surface due to noise and low resolution in the distance image.) Note that, in general, it is not possible to maintain equal distances between the grid cells when they are snapped to an arbitrary curved surface.
Therefore an algorithm which simulates the behavior of the physical grid patch is used to minimize the amount of stretching and bending during the deformation.
5. Interpolate between the grid cells to compute the grid vertices. Figure 12E
illustrates a deformation result.
[00821 Figures 13A-13C illustrate images associated with a series of operations that can be carried out for frame marker segmentation. Figures 9 and 10 illustrate an example of a frame 50f with fiducial markers 50fm circumferentially spaced apart around a center "cp" and around a space for the burr hole with two of the markers 50fmt, 50fm2 closer together than a third 50fm3 (Figure 10). Different arrangements, configurations or numbers of fiducial makers as well as different locations on the frame 50f and/or guide 50t may be used and the frame segmentation can be altered accordingly.
[00831 Some or all of these operations illustrated in or described with respect to Figures 13A-C may be carried out behind the scene (e.g., not actually displayed). Figure 13A illustrates an exemplary histogram with a noise region.
Figure 13B illustrates a cross-section of an imaged frame-marker. Figure 13C
illustrates a result of fitting a circle to the edge mask shown in Figure 13B.
The frame marker segmentation can define the orientation of the trajectory guide on the patient that can be correlated to mechanical output to cause the trajectory guide 50t to translate to provide a desired trajectory, e.g., rotations and/or direction left, right, counterclockwise or clockwise or other translation for moving the actuators to change pitch and/or roll (or X-Y location).
[0084] With reference to Figure 13A, the amplitude of the background noise in the input image stack can be estimated. To do so, a histogram of the stack can be constructed, the first negative maximum of the slope of the histogram can be located, the first peak to the left of this maximum can be located (this is the noise peak). The difference between the noise peak and the first negative maximum of the slope can approximate the standard deviation of the noise. The noise threshold can be obtained using the Equation (1) above. Again, this can be considered a first step in segmentation. Then, Steps 2-6 can be carried out with respect to the frame marker segmentation based on fit of an expected fiducial geometry to the observed fiducial positions because the fiducial markers are arranged with a fixed geometric relationship inside the volume.
2. Use a region-growing algorithm to find all "clumps" of pixels that are above the noise threshold.
3. Discard all clumps whose volumes are far from the known frame-marker volume.
4. Discard all clumps whose bounding-box dimension are far from the known bounding-box dimensions of the frame-markers.
5. Among the remaining clumps, look for triplets whose spatial arrangement matches the known spatial arrangement of the frame-markers of the trajectory base 50b (i.e., the centroids of the clumps should form a triangle (e.g., an isosceles triangle) of known dimensions.) [0085] At this point, if the number of clumps found for each frame (where bilateral procedures are used, there are two frames) is not exactly 3, then the segmentation is deemed to have failed.
6. For each triplet of markers, refine the marker locations:
a. Extract a 2D image of each marker by reformatting the stack onto the plane defined by the 3 marker centroids.
b. Compute the Canny edge mask of each 2D marker image.
c. Fit a circle to the Canny edge mask. The circle diameter is set equal to the diameter of the physical frame markers. The fit is performed by moving the circle until its overlap with the edge mask is maximized. The center of the fit circle is taken to be the location of the frame-marker's center.
[0086] In some embodiments, circuit 30c can be configured so that the program application can have distinct ordered workflow steps that are organized into logical groups based on major divisions in the clinical workflow as shown in Table 1.
A user may return to previous workflow steps if desired. Subsequent workflow steps may be non-interactive if requisite steps have not been completed. The major workflow groups and steps can include the following features or steps in the general workflow steps of "start", "plan entry", "plan target", "navigate", and "refine,"
ultimately leading to delivering the therapy (here placing the stimulation lead).
TABLE 1: EXEMPLARY CLINICAL WORKFLOW GROUPS/STEPS
Grou S:te ~ Deseri Start Start Set overall procedure parameters (Optionally confirm hardware com atibi(it ACPC Acquire a volume and determine AC, PC, and MSP points Plan Target Define initial target point(s) for entry planning Entry Trajectory Explore potential trajectories to determine entry oint s Grid Locate physical entry point via fiducial grid.
ACPC With hole burred and frame attached, acquire a volume Plan and determine revised AC, PC, and MSP points.
Target Target Acquire high-resolution slabs (e.g., T2 slabs) to determine target positions in new volume.
Trajectory Review final planned trajectory prior to starting procedure.
Initiate Acquire slabs to locate initial position of canula.
Dynamically re-acquire scan showing position of top of Navigate Alignment canula. With each update show projected target position to determine when alignment is correct.
Insertion Acquire slabs as probe is inserted into brain. Verify that probe is following planned trajectory.
Target Acquire images with probes in place. Review position and redefine target if necessary.
Adjust XY Dynamically re-acquire scan showing position of bottom Offset of canula. With each update show projected target Refine position to determine when offset is correct.
Insertion Acquire slabs as probe is inserted into brain. Verify that probe is following planned trajectory.
Lead Once probe position is finalized, prompt user to place Placement DBS leads or other device.
Admin Admin Re ortin and Archive functionality.
[00871 TABLES 2A-2P provide additional examples of some exemplary operations that may be associated with exemplary workflow steps according to some embodiments of the present invention.
TABLE 2A: Workflow Group Start Step -Desc~i'` tio'n .
The start step can provide UI (User Interface) for selecting procedure laterality.
The start step can provide UI for selecting target: STN, Gpi, or Custom The start step can provide UI for specifying the scanner bore diameter or scanner type that defines the size.
The start step can provide UI for entering hardware identifier data (e.g., version code) from the disposable hardware kit. If the version does not match the version supported by the software, an error can be shown, and the application can remain disabled.
Table 2B: Plan Entry- AC-PC Step Descn tion While data is being sent to the application, the UI can be disabled.
Given a 3D MR series of a whole human head, the application can automatically identify the AC, PC, and MSP points.
The application can display reformatted coronal, sagittal, and axial views aligned to the current AC, PC, and MSP
points.
AC, PC, and MSP points can be editable in any MPR
view in this step. On changing these points, the views can update to realign to the new ACPC coordinate system.
If a new series is sent while in this step, it will replace the existing series and clear all annotations.
While detecting the AC, PC, and MSP points, the UI can be disabled.
If additional data belonging to the current 3D MR series is sent, the AC, PC, and MSP points can be re-calculated automaticall .
[0088] The AC, PC and MSP locations can be identified in any suitable manner. In some embodiments, the AC-PC step can have an automatic, electronic AC, PC MSP Identification Library. The AC, PC and MSP anatomical landmarks define an AC-PC coordinate system, e.g., a Talairach-Tournoux coordination system that can be useful for surgical planning. This library can be used to automatically identify the location of the landmarks. It can be provided as a dynamic linked library that a host application can interface through a set of Application Programming Interface (API) on Microsoft Windows . This library can receive a stack of MR
brain images and fully automatically locates the AC, PC and MSP. The success rate and accuracy can be optimized, and typically it takes a few seconds for the processing. The output is returned as 3D coordinates for AC and PC, and a third point that defines the MSP. This library is purely computation and is typically UI-less. This library can fit a known brain atlas to the MR brain dataset. The utility can be available in form of a dynamic linked library that a host application can interface through a set of Application Programming Interface (API) on Microsoft Windows .
The input to this library can contain the MR brain dataset and can communicate with applications or other servers that include a brain atlas or include a brain atlas (e.g., have an integrated brain atlas). The design can be independent of any particular atlas;
but one suitable atlas is the Cerefy atlas of brain anatomy (note: typically not included in the library). The library can be configured to perform segmentation of the brain and identify certain landmarks. The atlas can then be fitted in 3D to the dataset based on piecewise affine transformation. The output can be a list of vertices of the interested structures.
[0089] In some embodiments, the mid-sagittal plane (MSP) is approximated using several extracted axial slices from the lower part of the input volume, e.g., about 15 equally spaced slices. A brightness equalization can be applied to each slice and an edge mask from each slice can be created using a Canny algorithm. A symmetry axis can be found for each edge mask and identify the actual symmetry axis based on an iterative review and ranking or scoring of tentative symmetry axes. The ranking/scoring cam be based on whether a point on the Canny mask, reflected through the symmetry axis lands on the Canny mask (if so, this axes is scored for that slice). An active appearance model (AAM) can be applied to a brain stem in a reformatted input stack with the defined MSP to identify the AC and PC
points.
[0090] The MSP plane estimate can be refined as well as the AC and PC
points. The MSP plane estimate can be refined using a cropped image with a small region that surrounds a portion of the brain ventricle and an edge mask using a Canny algorithm. The symmetry axis on this edge mask if found following the procedure described above. The AC and PC points are estimated as noted above using the refined MSP and brightness peaks in a small region (e.g., 6x6 mm) around the estimate are searched. The largest peak is the AC point. The PC point can be refined using the PC estimate above and the refined MSP. A Canny edge map of the MSP
image can be computed. Again, a small region (e.g., about 6 mm x 6mm) can be searched for a point that lies on a Canny edge and for which the image gradient is most nearly parallel to the AC-PC direction. The point is moved about 1 mm along the AC-PC direction, towards PC. The largest intensity peak in the direction perpendicular to AC-PC is taken to be the PC point.
Table 2C: Plan Entry - Target Step D;escrr n J
The application can provide the ability to save position coordinates as default values for the STN and Gpi targets. Initially these default values are set to 0,0,0.
Saved values can appear as the default for subsequent procedures using that target.
The application can also provide custom targets for which no default coordinate is supplied. When this option is selected, the user will be able to define a set of custom-named targets associated with a single entry point.
The application can display anatomic coronal, sagittal, and axial views.
Target points can be editable in any MPR view in this step.
The application can provide functionality to automatically register a brain atlas to the patient volume and generate outlines of structures associated with the selected target and display corresponding contours on the MPR views.
The application can provide interface to manually scale and offset the brain atlas registration to better match observed patient anatomy.
Table 2D: Plan Entry - Trajectory Step ., ;
For each given target point, the user can specify the corresponding entry point.
The application can display oblique reformatted coronal, sagittal, and axial views aligned to the proposed trajectory.
Entry points can be editable on either the oblique sagittal or coronal viewports in this step.
The oblique axial view can provide cine functionality to animate a fl -throu h along the trajectory.
The application can also provide an anatomical axial view.
The respective positions of the anatomical and oblique axial views can be represented by lines on the oblique sa ittal and coronal views.
When multiple targets (custom targets) have been defined for an entry point, the application can provide means to select the current target to display. Edits to the entry point will change the entry point for all associated tar ets.
If the user attempts to set the trajectory such that the probe could not be inserted without striking the bore an error can be shown and the trajectory will not be set.
This makes use of the bore size typically entered on the Start step.
The application can provide means to define named trajectories within the step. Trajectories from the list of named trajectories may be selected for display in the step.
If the user moves the entry point off the edge of the grid, warnin text will be shown.
Table 2E: Plan Entry - Grid Step ~'~~esCrl _t10~1 e ?~~~ ~~~~ ?NWW~ ?1 The application can display a volumetric 3D view showing the planning volume. For a bilateral procedure 2 such views can be shown, the left side in the left viewport, the right in the right viewport.
The application can optionally display the grid coordinates of the marking grid.
The application can optionally display overlay graphics to -visually identify edges and positions within grid squares.
On each view, the application can display the corres ondin entry point The application can automatically align the 3D views such that the user's point of view is looking along the trajectory from the entry towards the target. In the case of multiple custom targets, the trajectory to the first target can be used.
The application can set the visualization parameters of the volume such that the grid itself is visible to the user.
The initial zoom level can ensure that the entire head is visible.
100911 At this point, holes have been burred at the entry points and the trajectory guides 50t have been attached. NOTE: Because the patient has been moved, points defined in the previous image coordinate system may no longer be valid.
Also, brain shift may occur at this point.
Table 2F: Plan Target - AC-PC Step Descri pti The Plan Target AC-PC step can look and function the same as in the planning AC-PC step. However, data received in this step can be stored as the replanning volume. The AC, PC, and MSP annotations and the resultant transformation derived from the replanning volume can be kept distinct from those determined in the planning AC-PC step.
Table 2G: Plan Target- Target Step , ,. .
The Plan Target Step can function the same as in the planning step but with additional functionality to support slab data fusion.
The Plan Target Step can accept DICOM slab data.
While receiving slab data, the UI can be disabled and a message should be shown to indicate that a data transfer is in ro ress.
The Plan Target Step can provide a thumbnail bar that lists series in their order of acquisition. Selecting a series in the thumbnail bar will cause it to appear fused with the re-planning volume. Selecting the Plan Target volume can cause it to be dis la ed by itself.
Fused data can appear in the viewports along with the -plan target volume images and will be positioned and scaled to exactly coincide with the position and scale of the plan target volume.
The control panel can contain a slider that controls the relative intensity of the two series in the blended viewports.
The step can display scan plane parameters for an anatomical axial slab through the current target.
Table 2H: Plan Target - Trajectory Review Step Re,, uireme.nt des'cn tion The Plan Target Trajectory Review Step can function the same as the Planning Trajectory review step with only the following exceptions:
- Slab fusion support - Segment out the pivot point from frame markers - Use pivot point position as a fixed entry point (not editable) The Plan Target Trajectory Step can accept DICOM slab data.
While receiving slab data, the Ul can be disabled.
The Plan Target Trajectory Review Step can provide a thumbnail bar that lists series in their order of acquisition.
Selecting a series in the thumbnail bar will cause it to appear fused with the re-planning volume. Selecting the re-planning volume can cause it to be dis la ed by itself.
Fused data will appear in the viewports along with the volume images and will be positioned and scaled to exactly coincide with the position and scale of the re-planning volume.
The control panel can contain a slider that controls the relative intensity of each series in the blended viewports.
If for any reason the software is unable to identify the frame markers to find the pivot point, a warning can be dis ia ed.
The step can display scan plane parameters for:
- an oblique sagittal slab along the trajectory - an oblique coronal slab along the trajectory The step can display the trajectory angles relative to the anatomical coronal and sagittal planes.
Table 21: Navigate - Initiate ;Descr~ t_~on .
The step can prompt the user to acquire a small high-resolution slab through the proximal canula at a distance such that it will show a cross-section of the proximal canula even at maximum angulation and maximum offset. The slab can have a minimum of 4 slices.
(Example: given a canula 83 mm long, a maximum angulation of +/- 35 degrees, and a maximum offset of 4mm, then a scan plane 65 mm up from the distal canula marker will be sufficient to ensure that the canula is visible in the slab.) The application can also prompt the user to acquire a small high resolution slab scan with the following attributes:
= plane aligned to the plane of the frame markers (this can be based on the frame segmentation that was done for the trajectory review step) = plane center is positioned at the mechanical center of rotation = slab thickness is large enough to include all of the distal canula marker even under maximum angulation and offset (Example: given a maximum angulation +1-35 degrees, a maximum offset of 4 mm, and distal canula marker size of 7.1 mm, the total thickness required would be 11.6 mm, so any larger value may be used, say 13 mm) = slice spacing can be about 1 mm The application can identify the positions of the proximal and distal canula.
Using the detected positions of the frame markers in the plan target volume, the application can compare the observed position of the distal canula with the mechanical center of rotation. Since a locking pin may be used to ensure that there is no offset, values above a low threshold can cause a warning to be displayed.
(Example warning text: "Distal canula marker not found at expected location. Verify that canula is locked in `down' osition and reacquire distal canula scan."
If no pivot point marker can be identified, the user will be prompted to verify that the canula is locked in the 'down' position and re-acquire the slab scan.
The step can provide 3 MPR viewports in which to display the acquired slabs. These viewports will be oriented such that their base planes are aligned to the detected canula axis.
The step can provide a thumbnail bar to allow the user to select which ac uired slab to dis la .
Table 2J: Navigate - Alignment Descr,-i = tion The application can prompt for an alignment scan with the following attributes:
= scan plane is perpendicular to planned trajectory = scan plane is centered around the trajectory = scan plane position is set such that a cross-section of the proximal canula will be shown even at maximum angulation and maximum offset.
= a single 2D image can be acquired The application can display an anatomical axial view through the currently selected target.
The user may opt to switch the display to show a tra'ecto -axial view.
On receiving a 2D image of the top of the canula, the application can automatically identify the position of the top of the canula in 3D space. Using this position and, previously-determined pivot point, and the previously-determine offset, the application can draw an annotation representing the intersection of the current trajectory with the image plane containing the plahned target.
This step can display lines from the current projected target to the planned target that indicate the track the projected target would travel if the pitch and roll wheels were turned independently. These lines can be colored to match colors on the control wheels for pitch and roll respectively. A tool-tip (e.g., pop-up) can provide text to describe the suggested action.
For example: "Turn Roll knob to the Left"
On re-calculating the projected target point, an error value can indicate the in-plane linear distance between the projected target point and the planned target point on the currentl -dis la ed plane.
Images that are not oriented correctly to the requested scan plane can result in a warning. In this case, annotations and error measurements may not be dis la ed.
Images in which the targeting canula cannot be identified can result in a warning. Annotations and error measurements may not be dis la ed.
When multiple targets have been defined for an entry point, the application can provide means to select the current trajectory to dis la .
When drawing the target and the current projection of the canula path, the annotations can be drawn to match the physical size of the probe diameter.
Table 2K: Navigate - Insertion Desc:n tion The application can provide a depth value to set on the depth stop prior to insertion.
The application can prompt with scan parameters for oblique coronal and sagittal planes aligned to the trajectory. Also for an oblique axial perpendicular to the tra'ecto .
On receiving coronal or sagittal images, the application can display an overlay graphic indicating the planned trajectory. The most recent coronal and sagittal images can appear together in a 1 x2 dis la .
On receiving a trajectory axial scan perpendicular to the trajectory, the application can segment out the cross-sections of the probe to determine the actual path being followed by the probe.
On receiving a trajectory axial scan perpendicular to the trajectory, the application can display two viewports containing:
= the axial stack with graphic overlays showing the detected path of the probe on each image = an anatomic axial view through the target showing the planned target and the target projected from the detected path of the probe. An error value can show the distance between the current projected target and the planned target.
If multiple trajectories have been defined for a single entry, the application can display the trajectory that is currently aligned during insertion.
On entering the insertion step, the application can instruct the user to ensure that if they are using an imaging probe that it is connected to the scanner as an internal coil. Failure to do so could cause heating of the coil and injury to the patient. The user must explicitly click a button to acknowledge that they understand the warning.
Table 2L: Refine - Target The step can prompt for either i) a high-resolution 2D
image to be acquired using the imaging probe or ii) a high-resolution slab through the target area. The associated scan plane parameters can specify a trajectory axial image centered on the target.
The application can provide means to identify the tip of the probe (or stylet), and provide an error value for the linear distance from the probe to the planned target point in the axial anatomic plane.
The application can provide UI to set an updated target point.
The user may opt to proceed to the X-Y Adjustment step, return to Alignment to align to another target, or advance to the Admin step.
On accepting the current position, the user can be shown a warning not to scan once the MR-incompatible DBS
leads (if they are incompatible or potentially incompatible) have been placed. For bilateral cases, the user proceeds to complete the insertion of the probe on both sides before placing the leads. The message can indicate that scanning with MR-incompatible leads may result in serious injury or death.
If the user modifies a target point, the step can prompt the user to confirm removal of the offset locking pin. from the targeting frame before going on to the next step.
[0092] In the event that the placement is not acceptable, the user may opt to proceed to the X-Y Adjustment workflow step.
Table 2M: Refine - Adjust X-Y Offset ~ D~esc,yi tion The X-Y Adjustment step can display the current target and projected point as annotations to the image data that was acquired during the Target Refinement step.
This step can prompt the user to acquire 2D images with scan plane parameters such that the image lies perpendicular to the trajectory and through the pivot oint.
On receiving a 2D image through the pivot point, the step can calculate the current projected target and display an annotation on the 2D image from the imaging probe.
This step can display lines from the current projected target to the revised target that indicate the track the projected target would travel if the X and Y offset wheels were turned independently. The lines can be colored to match colors on the control wheels for X and Y offset respectively. A tool-tip (e.g., pop-up) can provide text to describe the necessary action.
(For example: "Turn X-offset knob to the Left"
This step can display an annotation indicating the location of the original planned target.
When drawing the target and the current projection of the canula path, the annotations can be drawn to match the physical size of the probe diameter.
Table 2N: Refine - Insertion Desc,r~ .tion .
After completing the X-Y Adjustment, the application can provide a workflow step to guide insertion. This is substantially the same as the first instance of the Insertion step above.
Table 20: Refine - Lead Placement , ;~ . .. _., D;esc After probe has been placed (or both probes for bilateral case) and position has been accepted by user, the user may proceed to the lead placementstep.
This step can provide the user with the depth values for the placement of the leads.
This step may advise the user that once leads have been placed scans may not be performed because heating in the leads could cause injury or death to the patient.
Table 2P: Admin Step Descn:t~on The step can provide means to archive data relating to the procedure. This includes:
= trajectory planning data = log= files with case data = image data The step can provide functionality to automatically generate a report documenting the performed procedure.
This report can include:
= patient information = AC, PC, and MSP points in MR space (Both detected and user-specified, if user modified.) = Planned and corrected targets in both MR and ACPC space = elapsed time for the procedure = physician case notes (optional) = any screenshots taken during the procedure An anonymous version of the report can also be generated automatically with the patient name and id removed.
The step can provide UI whereby the user can:
= selected a target = specify a position in MR coordinates representing the lead tip = define a set of offsets indicating the electrode offsets from the lead tip For each offset value, the step can find provide the ACPC coordinate that corresponds to a point offset from the tip position back along the trajectory of the lead.
These values correspond to the electrode positions in ACPC space. These values may be added to a patient report.
The Admin step can include a button to shut down the application on completion of the procedure. The user may not be allowed to otherwise close the application.
The application can have a configuration value to specify whether all patient data is to be cleared from the system on shut down.
Again, it is noted that functional patient data can be obtained in real-time and provided to the circuit 30c/workstation 30 on the display 32 with the visualizations of the patient anatomy to help in refining or planning a trajectory and/or target location for a surgical procedure.
[0093] When displaying images or visualizations that were created with the imaging probe 50a (Figure 5, where used), the circuit 30c may electronically apply a (sigma) correction to correct for a`volcano' or 'halo' shaped intensity distortion. That is, in some particular embodiments, the imaging probe antenna or coil may introduce a distortion in the images that are created with it that may be described as a bright halo around the probe itself where it appears in the image. Thus, when images from the imaging probe have such a feature, the circuit 30c can be configured to electronically automatically apply a correction to cancel out the halo (or at least reduce it). This feature will only affect the small field, high-resolution images that are created via the imaging probe itself. Images obtained using the main head coil do not typically have such a distortion.
10094] Referring to Figure 14A, the circuit 30c can be configured with a single control tool bar 30t that is displayed on the display 32 that allows the user to select what group and step to go to and also shows which groups and steps have been completed. Figure 14B shows an example of the workflow control tool bar 30t with the "Plan Entry" group selected, and the "Define Target" enlarged as the current step.
The tool bar 30t can include a color border 130 that can be used to partially or totally surround a button 130b to illustrate completion of a step. For bilateral procedures, the border can be color enhanced on one side when a task for that side is completed, e.g., on the left side when the left target is complete in the Define Target step.
[0095] As the user works through the procedure, certain clinical information is stored to be incorporated into a procedure report that may be reviewed at the end of the procedure and/or archived as a patient record. The circuit 30c can be configured to provide a user interface (UI) 301 that provides viewing tools, such as one or more of the following features.
- Draw measurement lines - Pan view - Zoom, Zoom All, Zoom to Region, Zoom to Point - Magnifying glass - Show/Hide Annotations - Show/Hide Crosshairs - Drag views between panels - Resize view panels - Maximize view to a lxl display - Save screen capture (can be added to the report) - Reset view settings to default [0096] Figure 15 is a screen shot of an exemplary UI 301 for the Start Group which may conveniently be configured as a one-screen input to set overall procedure parameters such as laterality, target type and MR Scanner bore size (recognizing that open bore MRI Scanner systems may also be used). Instead of bore size, a drop-down list can be provided that allows a user to select a manufacturer and type of MR Scanner in use that provides the associated bore size. Of course, the system 10 can be configured as an MR Scanner-specific system or the MR
Scanners of the future may have a standard bore size or be configured so that bore size is not a constraint and this information may not be required. As shown, the UI 301 can also include an input 321 that requires entry of hardware identifier data 66, shown as disposable kit version input, as discussed above with respect to Figure 5. In order to assure hardware and software compatibility and/or proper operation, if the identifier data 66 does not match, the system 10 can be configured to not allow a user to proceed to a next step or may prompt the user for other key codes.
[0097] Bore size is used in the step Plan Entry / Trajectory. If the user selects a trajectory such that the probe cannot be inserted into the canula 60 because it will not physically fit inside the scanner bore, a warning is generated (visual on the display 32 and/or audible). See Plan Entry / Trajectory step above.
[0098] Figure 16 is a screen shot associated with an exemplary UI 301 with a Plan Entry workflow group for a Define Target step (shown as left STN) illustrating the toolbar 30t and cross hair location data with default landmarks with patient image data. This step can be used to establish the AC PC coordinate system.
On receiving a whole-head volume, the step can automatically find candidates for these points. The user is to review and correct these points if necessary. The user can either position the crosshairs at a point and "click" the `Set' button to set=
the desired annotation, or they can drag an existing annotation around on the screen. Once points are defined, view planes can automatically reformat to align them to the ACPC
coordinate system to show the anatomical planes: Coronal, Sagittal, and Axial.
Any subsequent edit to the landmarks can cause the view planes to instantly re-align to match.
[0099J Figure 17 illustrates another exemplary screen shot for a Define Target step which is used to set target points so that the trajectories through potential entry points can be investigated in the next step. The user may opt to overlay the outlines from a standard brain atlas over the patient anatomy for comparison purposes which may b provided in color with different colors for different structures.
Figure 17 shows the play entry of the Define Target step with no atlas. Figure 18 illustrates the UI with an atlas showing a target outline in three orthogonal views and Figure 19 illustrates an atlas showing structures in a 1 x 1 layout. The view planes show the anatomical planes as defined in the ACPC step. Target points are edited similarly to how the ACPC points are edited. For the bilateral case, once the target has been defined for one side, then when the user selects the target for the other side, the crosshairs will automatically jump to the mirror-image position. If the patient has symmetric anatomy, this will save time in finding the equivalent position.
When using the brain atlas, the user may opt to show either just the target structure (STN or GPi) or all structures. In either case, a tooltip (e.g., pop-up) can help the user to identify unfamiliar structures. The user may also opt to scale and/or shift the brain atlas relative to the patient image to make a better match. To do this, the user may drag the white outline surrounding the brain atlas template. Fiber track structures and/or functional information of a patient's brain can be provided in a visually prominent manner (e.g., color coded or other visual presentation) for a surgeon's ease of reference. This information can also be selected or suppressed from views via a UI
selection 32F (e.g., toolbar option) as shown in Figure 4.
[01001 Figure 20 is a screen shot of an exemplary screen display for a Trajectory step in the Plan Entry workflow group. This step is used to find a clinically viable trajectory that determines the entry point on the skull typically via a grid such as that provided by the grid patch 50g (Figure 11). A visual warning can be displayed, e.g., a red warning message 30W on the top of the two top views and a red trajectory line 30R) can be used to indicate that the selected trajectory does not intersect the grid. In operation, upon entering this step, the workstation 30 can automatically search through the image volume for the marking grid(s) 50g. It can be configured to position the initial trajectory such that it runs through the middle of the grid 50g. If the user moves the trajectory such that the entry point is not on the grid, a warning 30W is displayed.
[0101] The top two views of Figure 20 show the coronal and sagittal views aligned along the current trajectory line. The user drags the trajectory line and it rotates freely about the target point. The bottom-left view shows the plane perpendicular to the trajectory at the level of the green line in the coronal and sagittal views. This is the "probe's eye" view. To advance the probe's eye view along the trajectory, in addition to dragging the green line along the trajectory line or using the mousewheel, onscreen VCR-style controls can provide an animated fly-through.
[0102] Figure 21 illustrates a (pop-up) warning 30W' that is automatically generated when a user selects a trajectory that may be blocked by the scanner bore wall. That is, if the user sets a trajectory such that the scanner bore will interfere with the insertion of the probe, a warning is displayed. This calculation is based on the bore size, current trajectory angles, and pre-configured values for the size of the probe 50a/50s/52, canula 60, and frame/trajectory guide 50t (Figure 5). Similar to the target step, in the bilateral case once the trajectory has been set on one side, the initial default for the other side can be a mirror-image trajectory to start the user closer to a more a likely trajectory. On saving the trajectory, the step automatically finds the surface of the skull along the planned/selected trajectory, identifies the coordinates on the grid and stores that location as the entry point.
[0103] Figure 22 illustrates an exemplary grid 50g shown overlying a patient's skull on a display 32 (without annotation lines), illustrating coordinates for selecting an entry location (shown as columns 1-6 and rows A-F) with a left STN
entry location. Figure 23 shows the same screen view as Figure 22 but with optional annotation grid lines. Typically, a user can see the grid coordinates clearly enough that the optional overlay grid lines are not required in order to identify the grid elements. However, in some embodiments, as shown a UI will allow a user to display the lines.
[0104] Figure 24 is an exemplary screen shot of a Plan Target Group workflow with the ACPC step selected. This workflow group is used to precisely determine target points after one burr hole has been formed for a unilateral procedure or both burr-holes have been formed for bilateral procedures (the surgical entries have been burred and the frames attached). Previous planning is typically invalidated by brain shift that occurs with the loss of cerebral spinal fluid pressure. This step is substantially similar to the Plan Entry ACPC step discussed above. The difference is that in addition to the whole-head volume, the user may optionally also send one or more high-resolution slabs containing the ACPC points. The user can use a thumbnail bar to select which data to use to display and edit the ACPC landmarks. Figure illustrates a screen shot with a Plan Target/ACPC step showing slab data.
[0105] Figure 26 is an exemplary screen shot of a Plan Target Group workflow with the Define Target step selected. The brain images can be shown with blended volume and slab image data. This step is similar to the Define Target step in the Plan Entry group, but like the preceding step it also supports high-resolution slabs to increase the precision with which the target can be defined. Thus, this step has the ability to show a blended image using data from both the volume and a high-resolution slab. A slider 30S (shown to right under the word "blend") in the UI allows the user to control the ratio of how much of the image comes from either source.
[0106] Figure 27 is a screen shot of a Trajectory step in the Plan Target workflow group. This step is similar to the Plan Entry/Trajectory step, but in this case this step is typically only used to review the trajectory. The entry point cannot be changed since the hole has been burred and the frame attached. Also, like the preceding step, this can support the use of high-resolution slabs and can blend between the volume and the slab. On entering this step, the software can automatically search through the whole-head volume and find the frame marker fiducials. From these, the system 10 and/or circuit 30c can determine the frame locations and orientation and uses this to calculate the actual entry point on the head.
This is used along with the target from the previous step to determine the trajectory.
Otherwise, the trajectory may be reviewed as described above in the Plan Entry /
Trajectory step.
[0107] Figure 28 is a screen shot of a Navigate workflow group with the Initiate Navigation step selected. This visualization shows the proximal canula slab.
On entering this group, the user has already burred the entry hole, attached the frame, and finalized their planned targets. This group will guide the user through aligning the canula 60 (Figure 8) to match the planned trajectory and inserting the probe through the canula to the correct depth. This step is used to determine the initial physical location of the canula. The user acquires a scan through the top of the canula and another through the bottom. The circuit 30c (e.g., software) automatically finds the canula in each slab and determines its position and orientation in space.
Figure 29 illustrates the distal canula slab in the Navigate/Initiate Navigation step.
[0108] Figure 30 is a screen shot of a Navigate workflow group with the Align Canula step selected. This step is used to physically align the canula .5 (Figure 8) to the planned trajectory. The user iteratively adjusts the canula angulation via the pitch and roll control knobs as they rapidly re-acquire an image through the top of the canula 60p. With each update, the circuit 30c (e.g., software) calculates the position of the canula 60 and displays an annotation showing where it's currently pointing on the target plane. Figure 30 illustrates that user feedback such as a prompt over overlay or a tooltip-style pop-up can tell the user which control to turn and which direction. In some embodiments, as shown, annotations can be drawn as circles corresponding to the probe diameter. The circle with the crosshair is the planned target, the other circle is the current projected point based on the trajectory of the canula. The lines in between show the relative amount of pitch and roll to apply and the text can specify which known to turn and in which direction (shown as "turn pitch wheel to the left").
[0109] Figure 31 illustrates a screen shot from an exemplary Navigate/Insert Probe step. This screen shot illustrates coronal and sagittal views to the target (e.g., STN) and can provide a set depth stop dimension (shown on the upper right hand side of the UI). This step allows the user to see how well the probe is following the trajectory as it is inserted. The user may opt to scan Coronal and Sagittal slabs along the probe to visually determine the probe alignment in those planes. The user can also scan perpendicular to the probe. In that case, the circuit 30c (e.g., software) can automatically identify where the probe is in the slab and it then shows a projection of the current path onto the target plane to indicate the degree and direction of error if the current path is continued. Figure 32 illustrates an axial slab and projected point with a projected error if the current trajectory continues (the probe is shown in the right image offset by 4.6 mm and the offset is also noted on the UI as "Projected Error". The user can perform these scans multiple times during the insertion. The automatic segmentation of the probe and the display of the projected target on the target plane provide fully-automatic support for verifying the current path. The Coronal/Sagittal views can provide the physician with a visual confirmation of the probe path that doesn't depend on software segmentation.
[01101 Figure 33 illustrates an exemplary screen shot of a Refine Placement workflow group with the Target Revision step selected. This step can illustrate the target slab. After completing the initial insertion, the user (e.g., physician) may find that either the placement doesn't correspond sufficiently close or perfectly to the plan, or the plan was not correct. This may be particularly likely if an imaging probe (50a) is used, since a user will be able to more clearly visualize structures like the STN that are usually indistinct with external coils. This workflow group can support functionality whereby the physician can withdraw the probe and use the X and Y offset adjustments to obtain a parallel trajectory to a revised target.
This step can prompt the user or otherwise acquire an image slab through the distal tip of the probe. (Optionally, this step may use the imaging probe). The step displays the slab and on it the user may opt to modify the target point to a new location or accept the current position as final.
[0111] Figure 34 illustrates an exemplary screen shot of the Refine Placement workflow group with the Adjust X-Y Offset step selected. This step is very similar to the Navigate/Align Canula step described above. The primary difference is that instead of adjusting the angulation of the canula 60, the user is adjusting a small X-Y offset to set the canula 60 to a trajectory parallel to the original one. Figure 34 shows the display with an visualization f the position of the probe tip relative to the target and with instructions on what physical adjustments to make to obtain the desired parallel trajectory (shown as "turn Y wheel to the right") and the projected error (shown a 2.7 mm). Figure 35 illustrates a detail of the adjust annotations and pop-up (shown as "turn X wheel to the left").
[0112] After the X-Y adjustments are made, the Insert Probe of the Refine Placement workflow group is selected and carried out in the same manner as the Navigate/Insert Probe step described above.
[0113] Figure 36 illustrates the Refine Placement workflow group with the "conclude procedure" step selected. This step occurs after all probes have'been inserted and have had their positions verified by the physician. At this point, the U1 can prompt them to insert both leads where implantable stimulation leads are to be placed (using the defined trajectory) and can warn them not to perform any additional scans if MRI-incompatible or potentially incompatible leads are used. As shown, the system 10 can be configured to define (and output to a user) or depth stops to set the lead or other therapy or diagnostic device for each STN or target site. The depth stops can be different for each implant location on the left and right targets (for bilateral procedures) so that the electrodes of the leads or other components for other devices are positioned in the desired location.
[0114] Figure 37 illustrates an example of a display that may be used for the ADMIN workflow group. This group has one step that provides reporting and archive functionality. The report automatically documents the entire procedure including annotations, measurements, and screen captures. The circuit 30c can generate a full version and an anonymous version of the report and may include a date as to when everything is archived to CD.
[0115] The circuit 30c may also be configured to determine where individual electrodes on the DBS leads are situated in ACPC coordinates. Given the tip position in MR coordinates (the circuit 30c can ill in the planned position, but the user may change it) the user will provide a set of offset values that represent the distance of each electrode from the lead tip. In other embodiments, a lead type can be selected such as from a pull-down list and those values can automatically be input based on the manufacturer and lead type (e.g., design thereof). The circuit 30c can be configured so that the UI displays the corresponding electrode positions in ACPC
coordinates.
[0116] Figure 38 illustrates an example of a display with a UI that may be used for the ADMIN workflow group shown as- Admin Page/Electrode Offset dialog step selected. This step may also be included in the Conclude Procedure/step or provided as a separate workflow group. The electrode offset values may significantly speed up the process by which the pulse generator is programmed since the physician will know where the electrodes are anatomically.
[0117] The system 10 may also include a decoupling/tuning circuit that allows the system to cooperate with an MRI scanner 20 and filters and the like. See, e.g., U.S. Patent Nos. 6,701,176; 6, 904,307 and U.S. Patent Application Publication No. 2003/0050557, the contents of which are hereby incorporated by reference as if recited in full herein. As noted above, one or more of the tools can include an intrabody MRI antenna 50a (Figure 5) that is configured to pick-up MRI signals in local tissue during an MRI procedure. The MRI antenna can be configured to reside on the distal portion of the probe. In some embodiments, the antenna has a focal length or signal-receiving length of between about 1-5 cm, and typically is configured to have a viewing length to receive MRI signals from local tissue of between about 1-2.5 cm. The MRI antenna can be formed as comprising a coaxial and/or triaxial antenna. However, other antenna configurations can be used, such as, for example, a whip antenna, a coil antenna, a loopless antenna, and/or a looped antenna.
See, e.g., U.S. Patent Nos. 5,699,801; 5,928,145; 6,263,229; 6,606,513; 6,628,980;
6,284,971;
6,675,033; and 6,701,176, the contents of which are hereby incorporated by reference as if recited in full herein. See also U.S. Patent Application Publication Nos.
2003/0050557; 2004/0046557; and 2003/0028095, the contents of which are also hereby incorporated by reference as if recited in full herein.
(0118] In some embodiments, the implanted leads and/or intrabody tools can be configured to allow for safe MRI operation so as to reduce the likelihood of undesired deposition of current or voltage in tissue. The leads or tools can include RF
chokes such as a series of axially spaced apart Balun circuits or other suitable circuit configurations. See, e.g., U.S. Patent No. 6,284,971, the contents of which are hereby incorporated by reference as if recited in full herein, for additional description of RF
inhibiting coaxial cable that can inhibit RF induced current. The conductors connecting electrodes or other components on or in the tools can also include a series of back and forth segments (e.g., the lead can turn on itself in a lengthwise direction a number of times along its length) and/or include high impedance circuits. See, e.g., U.S. Patent Application Serial Nos. 11/417,594; 12/047,602; and 12/090,583, the contents of which are hereby incorporated by reference as if recited in full herein.
10119] Although not shown, in some embodiments, one or more of the surgical tools can be configured with one or more lumens and exit ports that deliver desired cellular, biological, and/or drug therapeutics to the target area, such as the brain. The tools may also incorporate transseptal needles, biopsy and/or injection needles as well as ablation means. The lumens, where used, may receive extendable needles that may exit the probe from the distal end or from the sides, proximal, distal, or even, through the electrodes to precisely deliver cellular/biological therapeutics to the desired anatomy target. This delivery configuration may be a potential way to treat patients, where the cellular/biological therapeutics can be delivered into the desired anatomy to modify their cellular function. The cells (e.g., stem cells) may improve function. MRI can typically be effectively used to monitor the efficacy and/or delivery of the therapy.
[01201 The system 10 can include circuits and/modules that can comprise computer program code used to automatically or semi-automatically carry out operations to generate visualizations and provide output to a user to facilitate MRI-guided diagnostic and therapy procedures. Figure 39 is a schematic illustration of a circuit or data processing system that can be used with the system 10. The circuits and/or data processing systems may be incorporated in one or more digital signal processors in any suitable device or devices. As shown in Figure 39, the processor 410 communicates with an MRI scanner 20 and with memory 414 via an address/data bus 448. The processor 410 can be any commercially available or custom microprocessor. The memory 414 is representative of the overall hierarchy of memory devices containing the software and data used to implement the functionality of the data processing system. The memory 414 can include, but is not limited to, the following types of devices: cache, ROM, PROM, EPROM, EEPROM, flash memory, SRAM, and DRAM.
[0121] As shown in Figure 39 illustrates that the memory 414 may include several categories of software and data used in the data processing system:
the operating system 452; the application programs 454; the input/output (I/O) device drivers 458; and data 456. The data 456 can also include predefined characteristics of different surgical tools and patient image data 455. Figure 39 also illustrates the application programs 454 can include a Visualization Module 450, Interventional Tool Data Module 451, a Tool Segmentation Module 452 (such as segmentation modules for a grid patch, a targeting canula, and a trajectory guide frame and/or base), and a workflow group User Interface Module 453 (that facilitates user actions and provides guidance to obtain a desired trajectory such as physical adjustments to achieve same).
[01221 As will be appreciated by those of skill in the art, the operating systems 452 may be any operating system suitable for use with a data processing system, such as OS/2, AIX, DOS, OS/390 or System390 from International Business Machines Corporation, Armonk, NY, Windows CE, Windows NT, Windows95, Windows98, Windows2000 or other Windows versions from Microsoft Corporation, Redmond, WA, Unix or Linux or FreeBSD, Palm OS from Palm, Inc., Mac OS from Apple Computer, LabView, or proprietary operating systems. The I/O device drivers 458 typically include software routines accessed through the operating system 452 by the application programs 454 to communicate with devices such as I/0 data port(s), data storage 456 and certain memory 414 components. The application programs are illustrative of the programs that implement the various features of the data processing system and can include at least one application, which supports operations according to embodiments of the present invention. Finally, the data 456 represents the static and dynamic data used by the application programs 454, the operating system 452, the I/O device drivers 458, and other software programs that may reside in the memory 414.
[0123] While the present invention is illustrated, for example, with reference to the Modules 450-453 being application programs in Figure 39, as will be appreciated by those of skill in the art, other configurations may also be utilized while still benefiting from the teachings of the present invention. For example, the Modules 450-453 and/or may also be incorporated into the operating system 452, the I/O
device drivers 458 or other such logical division of the data processing system. Thus, the present invention should not be construed as limited to the configuration of Figure 39 which is intended to encompass any configuration capable of carrying out the operations described herein. Further, one or more of modules, i.e., Modules 450-453 can communicate with or be incorporated totally or partially in other components, such as a workstation, an MRI scanner, an interface device. Typically, the workstation 30 will include the modules 450-453 and the MR scanner with include a module that communicates wit the workstation 30 and can push image data thereto.
[0124] The I/O data port can be used to transfer information between the data processing system, the circuit 30c or workstation 30, the MRI scanner 20, and another computer system or a network (e.g., the Internet) or to other devices controlled by or in communication with the processor. These components may be conventional components such as those used in many conventional data processing systems, which may be configured in accordance with the present invention to operate as described herein.
[0125] In the drawings and specification, there have been disclosed embodiments of the invention and, although specific terms are employed, they are used in a generic and descriptive sense only and not for purposes of limitation, the scope of the invention being set forth in the following claims. Thus, the foregoing is illustrative of the present invention and is not to be construed as limiting thereof.
More particularly, the workflow steps may be carried out in a different manner, in a different order and/or with other workflow steps or may omit some or replace some workflow steps with other steps. Although a few exemplary embodiments of this invention have been described, those skilled in the art will readily appreciate that many modifications are possible in the exemplary embodiments without materially departing from the novel teachings and advantages of this invention.
Accordingly, all such modifications are intended to be included within the scope of this invention as defined in the claims. In the claims, means-plus-function clauses, where used, are intended to cover the structures described herein as performing the recited function and not only structural equivalents but also equivalent structures. Therefore, it is to be understood that the foregoing is illustrative of the present invention and is not to be construed as limited to the specific embodiments disclosed, and that modifications to the disclosed embodiments, as well as other embodiments, are intended to be included within the scope of the appended claims. The invention is defined by the following claims, with equivalents of the claims to be included therein.
Claims (49)
1. An MRI-guided surgical system, comprising:
at least one MRI-compatible surgical tool;
a circuit adapted to communicate with an MRI scanner, wherein the circuit electronically recognizes predefined physical characteristics of the at least one tool to automatically segment MR image data provided by the MRI scanner whereby the at least one tool constitutes a point of interface with the system; and at least one display in communication with the circuit, wherein the circuit is configured to provide a User Interface that defines workflow progression for an MRI-guided surgical procedure and allows a user to select steps in the workflow, and wherein the circuit is configured to generate multi-dimensional visualizations using predefined data of the at least one tool and data from MRI images of the patient in substantially real time during the surgical procedure.
at least one MRI-compatible surgical tool;
a circuit adapted to communicate with an MRI scanner, wherein the circuit electronically recognizes predefined physical characteristics of the at least one tool to automatically segment MR image data provided by the MRI scanner whereby the at least one tool constitutes a point of interface with the system; and at least one display in communication with the circuit, wherein the circuit is configured to provide a User Interface that defines workflow progression for an MRI-guided surgical procedure and allows a user to select steps in the workflow, and wherein the circuit is configured to generate multi-dimensional visualizations using predefined data of the at least one tool and data from MRI images of the patient in substantially real time during the surgical procedure.
2. An MRI-guided surgical system according to Claim 1, wherein the User Interface is configured to accept input of an identifier associated with the at least one tool and to block use of the system if the identifier of the at least one tool indicates that it is not an authorized tool or that the at least one tool has a version that is not compatible with the system.
3. An MRI-guided surgical system according to Claim 1, wherein the at least one tool is an imaging probe with an intrabody antenna that is configured to collect small field, high-resolution image data, and wherein the circuit is configured to electronically apply a correction to reduce intensity distortion associated with image data collected from the imaging probe.
4. A surgical system according to Claim 1, wherein the at least one tool comprises a flexible patch with a grid, and wherein the circuit is configured with a segmentation module that is able to locate the patch in an MR volume, and generate a visualization presented on the display which shows the patch as an overlay on a patient with defined grid coordinates for a surgical entry site.
5. A surgical system according to Claim 1, wherein the system is a brain surgery system, wherein the flexible patch is a brain surgery patch configured to reside on a patient's skull, and wherein the circuit is configured to identify the patch position and orientation on the patient based on predefined physical characteristics of the patch, electronically deform the patch to fit curvature of a patient's skull, and generate and display a visual overlay of the patch on the skull of the patient.
6. A surgical system according to Claim 1, wherein the system is a brain surgery system, wherein the at least one tool is a brain surgery flexible grid patch configured to reside on a patient's head, wherein the circuit is configured to visually illustrate on the display a user-selectable trajectory line to a deep brain location that intersects the grid and defines a location on the grid for marking a burr entry hole based on the desired trajectory line, and wherein the User Interface displays an audible and/or visual warning when the user selects trajectory line that does not intersect the grid.
7. A surgical system according to Claim 6, wherein the circuit is configured to display a location and/or coordinates associated with the grid that provides a desired burr hole location for a trajectory entry path through a skull of a patient.
8. A system according to Claim 1, wherein the at least one surgical tool comprises a trajectory guide with fiducial markers in a fixed geometric relationship, the trajectory guide configured to define a trajectory path for a subsurface brain target in the patient, the trajectory guide having a base that affixes to a patient's skull, the base having a substantially circular opening that aligns with a burr hole in the patient's skull and encloses a space for two pivot axes of rotation of the trajectory guide, and wherein the circuit is configured to segment MR image data to locate the fiducial markers of the trajectory guide and orient the trajectory guide.
9. A system according to Claim 8, wherein the trajectory guide fiducial markers includes a plurality of fiducial markers that surround the base opening, and wherein the circuit is configured to segment MR image data to locate the fiducial markers of the base and orient the trajectory guide.
10. A system according to Claim 9, wherein the fiducial markers are circumferentially spaced apart and positioned relative to each other such that at least two are closer together than another to define an affirmative orientation of the trajectory guide.
11. A system according to Claim 8, wherein. the fiducial markers comprise three fluid-filled annular fiducial markers.
12. A system according to Claim 8, wherein the circuit is configured to employ a statistical best-fit analysis to MRI image data to fit observed fiducial positions to expected fiducial geometry based on the fixed geometric relationship of the fiducial markers.
13. A system according to Claim 10, wherein the circuit is configured to fit observed signal intensity data to a known geometric circumferential relationship of the circumferentially spaced apart fiducial markers thereby providing for a more robust image recognition while inhibiting errors if one fiducial marker has less signal intensity than others.
14. A system according to Claim 8, wherein the circuit comprises a segmentation module for a base of a trajectory frame, a segmentation module for one or more markers associated with a targeting canula held by the trajectory guide, and wherein the circuit is configured, before a probe is inserted fully, to calculate an extrapolated projection along the probe's axis and display an error in position if a current insertion path is followed.
15. A system according to Claim 8, wherein the circuit is configured to calculate and provide suggested adjustments to carry out X-Y and/or pitch and roll changes for the trajectory guide to position a probe inserted through the targeting canula held by the trajectory guide in the brain at an identified target location
16. A system according to Claim 8, wherein the trajectory guide comprises a frame attached to an X-Y translating platform that resides on pitch and roll control arcs, an X adjustment actuator, a Y adjustment actuator, a pitch adjustment actuator and a roll adjustment actuator, each actuator associated with the platform to provide the associated adjustment to a pivot axis of the trajectory guide, wherein the circuit comprises a module that has predefined physical characteristics of the trajectory guide and is configured to electronically determined and output to a user via the display, which actuator to move and in which direction to align the trajectory to reach the target location.
17. A system according to Claim 8, wherein the circuit is configured to generate a visual of a selected projected trajectory path to a target site in the brain and display at least one of an associated X adjustment, Y adjustment, pitch adjustment, and roll adjustment for actuators associated with the trajectory guide to provide the selected trajectory path.
18. A system according to Claim 1, wherein the circuit has or is configured to receive data regarding a bore size of the MRI Scanner, and wherein the circuit is configured to monitor for a physical limitation or interference of a surgical tool based on: (a) pre-defined physical characteristics of the surgical tool; (b) MRI
scanner bore size; and (c) patient size, and wherein the circuit is configured to generate an audible and/or visual warning when the surgical tool will be blocked by physical interference with a wall defining at least a portion of the bore size of the MRI scanner.
scanner bore size; and (c) patient size, and wherein the circuit is configured to generate an audible and/or visual warning when the surgical tool will be blocked by physical interference with a wall defining at least a portion of the bore size of the MRI scanner.
19. A system according to Claim 1, further comprising a clinician workstation with the display in communication with the circuit, wherein the at least one tool comprises a trajectory guide with a frame that attaches to a patient, wherein the circuit is configured to segment MRI image data of the patient and define a roll axis and a pitch axis based on interrogation of the MRI data in substantially real-time to locate at least one frame marker and register the trajectory guide in position on the patient, and wherein the circuit is configured to calculate an insertion depth of a probe associated with the at least one tool to reach a selected intrabody target using the predefined physical characteristics of the trajectory guide.
20. A system according to Claim 1, wherein the circuit is configured to generate color coded electroanatomical visualizations of target anatomy in substantially real-time, using at least one of:
(a) electrical signals obtained from electrodes associated with the at least one tool, correlated to time and position of the electrodes to render visualizations to illustrate neural electrical potentials and locations; or (b) patient functional data including fMRI data and/or MRI image data showing electrical activity in response to a stimulation.
(a) electrical signals obtained from electrodes associated with the at least one tool, correlated to time and position of the electrodes to render visualizations to illustrate neural electrical potentials and locations; or (b) patient functional data including fMRI data and/or MRI image data showing electrical activity in response to a stimulation.
21. A system according to Claim 1, in combination with an MRI scanner, and further comprising a clinician workstation in communication with the MRI
scanner and the circuit, wherein the circuit comprises a DICOM interface that receives MRI
images from the MRI scanner, and wherein the MRI scanner comprises a module that allows communication with the clinician workstation.
scanner and the circuit, wherein the circuit comprises a DICOM interface that receives MRI
images from the MRI scanner, and wherein the MRI scanner comprises a module that allows communication with the clinician workstation.
22. A system according to Claim 21, wherein the circuit and clinician workstation passively receives data from the MRI scanner.
23. A system according to Claim 1, in combination with an MRI scanner, and further comprising a clinician workstation in communication with the MRI
scanner and the circuit, and wherein the circuit receives pre-DICOM reconstructed MR
image data and/or receives MR data and generates the image construction for the visualization.
scanner and the circuit, and wherein the circuit receives pre-DICOM reconstructed MR
image data and/or receives MR data and generates the image construction for the visualization.
24. A system according to Claim 1, wherein the circuit is configured to display substantially real-time patient functional information in the visualizations.
25. A system according to Claim 1, wherein the User Interface is configured with an input that allows a user to select whether to display patient functional data in substantially real-time in the visualizations and/or to display a patient's fiber tracks in the visualizations.
26. A method for carrying out an MRI-guided surgical procedure, comprising;
defining dimensional and/or functional data of at least one MRI compatible surgical tool;
obtaining MRI image data of the patient;
electronically segmenting the MRI image data to identify known fiducial markers on the at least one tool based on the defining step;
generating visualizations of the at least one tool registered to patient anatomical structure;
electronically generating directions on adjustments for a pitch, roll or X-Y
actuator to adjust a trajectory of a trajectory guide; and guiding the tool to a location in the patient using MRI image data, the directions for adjustment and the visualizations thereby facilitating an MRI-guided surgical procedure.
defining dimensional and/or functional data of at least one MRI compatible surgical tool;
obtaining MRI image data of the patient;
electronically segmenting the MRI image data to identify known fiducial markers on the at least one tool based on the defining step;
generating visualizations of the at least one tool registered to patient anatomical structure;
electronically generating directions on adjustments for a pitch, roll or X-Y
actuator to adjust a trajectory of a trajectory guide; and guiding the tool to a location in the patient using MRI image data, the directions for adjustment and the visualizations thereby facilitating an MRI-guided surgical procedure.
27. A method according to Claim 26, wherein the generating visualizations includes displaying patient function.
28. A method according to Claim 27, wherein the patient function is shown as active regions of a brain in response to a defined stimulation and/or fMRI in substantially real-time and/or displaying fiber tracks of the patient in the visualizations.
29. A computer program product for facilitating an MRI-guided surgical procedure, the computer program product comprising:
a computer readable storage medium having computer readable program code embodied in said medium, said computer-readable program code comprising:
computer readable program code that comprises predefined physical data of a plurality of different surgical tools;
computer readable program code that communicates with an MRI scanner to obtain MRI image data of a patient; and computer readable program code that generates visualizations of the patient using data from the tools and the image data of the patient in-substantially real-time.
a computer readable storage medium having computer readable program code embodied in said medium, said computer-readable program code comprising:
computer readable program code that comprises predefined physical data of a plurality of different surgical tools;
computer readable program code that communicates with an MRI scanner to obtain MRI image data of a patient; and computer readable program code that generates visualizations of the patient using data from the tools and the image data of the patient in-substantially real-time.
30. A computer program product according to Claim 29, further comprising computer readable program code that obtains image data of patient function, and wherein the computer readable program code that generates the visualizations also shows patient function including active regions in a brain based on fMRI
and/or patient stimulation.
and/or patient stimulation.
31. A MRI-guided interventional deep brain system, comprising:
an MRI Scanner;
a clinician workstation with a circuit and a display, the workstation in communication with the MRI Scanner;
a flexible patch with a grid thereon configured to releasably attach to a patient's skull; and a trajectory guide attachable to a skull of a patient, the guide having a base with an aperture configured to reside over a burr hole formed in a patient's skull, the base aperture providing a mechanical center of rotation for a pivot axis associated with the trajectory guide, the trajectory guide having a plurality of spaced apart fiducial markers;
wherein the circuit comprises physical data regarding the patch and is configured to interrogate patient imaging data provided by the MRI Scanner and segment the image data to define a burr hole location that intersects the patch with a desired intrabrain trajectory, and wherein the circuit comprises tool-specific data of the trajectory guide and is configured to interrogate patient imaging data provided by the MRI Scanner and interactively generate visualizations of the patient's brain and the trajectory guide to the display.
an MRI Scanner;
a clinician workstation with a circuit and a display, the workstation in communication with the MRI Scanner;
a flexible patch with a grid thereon configured to releasably attach to a patient's skull; and a trajectory guide attachable to a skull of a patient, the guide having a base with an aperture configured to reside over a burr hole formed in a patient's skull, the base aperture providing a mechanical center of rotation for a pivot axis associated with the trajectory guide, the trajectory guide having a plurality of spaced apart fiducial markers;
wherein the circuit comprises physical data regarding the patch and is configured to interrogate patient imaging data provided by the MRI Scanner and segment the image data to define a burr hole location that intersects the patch with a desired intrabrain trajectory, and wherein the circuit comprises tool-specific data of the trajectory guide and is configured to interrogate patient imaging data provided by the MRI Scanner and interactively generate visualizations of the patient's brain and the trajectory guide to the display.
32. A system according to Claim 31, wherein the circuit is configured to obtain patient function data and generate the visualizations showing patient function in substantially real time.
33. A system according to Claim 31, wherein the circuit is configured to provide a default trajectory for the trajectory guide on the display that extends through a center location of the grid patch.
34. A system according to Claim 31, further comprising at least one fiber optic device attached to the trajectory guide for viewing the burr hole with a local field of view, the fiber optic device being in communication with an MRI compatible camera.
35. A system according to Claim 31, wherein the system comprises two grid patches and two trajectory guides for a bilateral procedure, and two fiber optic video stream imaging devices, one mounted to each trajectory guide, each being in communication with a different MRI compatible camera mounted proximate a bore of the MRI Scanner, and wherein the workstation displays a video stream from one or both of the cameras while also displaying the visualizations.
36. A system according to Claim 31, wherein the system is configured to determine and display a depth value for a user to set on a depth stop associated with an intrabody probe or lead prior to insertion.
37. A system according to Claim 31, wherein the circuit is in communication with a User Interface for a display whereby a user can: (a) select an intrabrain target, (b) select or specify a position in MR coordinates representing a lead tip, and (c) select or specify an identifier that selects a set of offsets indicating electrode offsets from the lead tip, and wherein the circuit is configured, for each offset value, to provide an AC-PC coordinate that corresponds to a point offset from the tip position back along the trajectory of the lead whereby the AC-PC coordinates correspond to the electrode positions in AC-PC space.
38. A system according to Claim 31, wherein the circuit is in communication with a User Interface that includes a series of selectable workflow groups including "start", "plan entry", "plan target", "navigate", and "refine" that can be used to guide the surgical procedure resulting in delivering a therapy after the "refine"
workflow group.
workflow group.
39. A system according to Claim 38, wherein the User Interface allows a user to select whether to show patient function information in the visualizations.
40. A system according to Claim 38, wherein the User Interface also includes an "Administrative" workflow group whereby the circuit electronically generates a medical report automatically summarizing clinical information regarding the patient and certain surgical information including at least a plurality of the following:
(a) AC, PC, and MSP points in MR space (both detected and user-specified, if user modified);
(b) planned and corrected targets in both MR and ACPC space;
(c) elapsed time for the procedure;
(d) physician case notes (optional); and (e) screenshots taken during the procedure.
(a) AC, PC, and MSP points in MR space (both detected and user-specified, if user modified);
(b) planned and corrected targets in both MR and ACPC space;
(c) elapsed time for the procedure;
(d) physician case notes (optional); and (e) screenshots taken during the procedure.
41. A system according to Claim 31, wherein the circuit is configured to generate a warning when a user selects or draws a trajectory that does not intersect the grid patch.
42. A system according to Claim 31, wherein the circuit is configured to require hardware identifier data, and wherein if the identifier data is not recognized or is defined to be incompatible with the circuit, the system is configured to inhibit use of the system.
43. A system according to Claim 31, further comprising an imaging probe.
with an intrabody antenna that is configured to extend through the targeting canula and collect small field, high-resolution image data, and wherein the circuit is configured to electronically apply a correction to reduce intensity distortion associated with image data collected from the imaging probe.
with an intrabody antenna that is configured to extend through the targeting canula and collect small field, high-resolution image data, and wherein the circuit is configured to electronically apply a correction to reduce intensity distortion associated with image data collected from the imaging probe.
44. A system according to Claim 31, further comprising a User Interface that allows a user to select different intrabody procedures including a unilateral or bilateral procedure and a desired intrabody target, wherein if a bilateral procedure is selected, the User Interface provides a toolbar with left and right workflow steps.
45. A system according to Claim 44, wherein the User Interface for bilateral procedures comprises workflow steps in a visual control that guides a user to complete grid entry locations for both sides, burr hole formation for both sides and trajectory frame attachment to both sides before proceeding to a "plan target"
step due to brain shift and before directing a patient to be returned to an imaging location in the magnet.
step due to brain shift and before directing a patient to be returned to an imaging location in the magnet.
46. A system according to Claim 31, wherein the system comprises a User Interface with workflow steps, and wherein the User Interface comprises a control that allows a user to select to show patient function information in the visualizations in substantially real-time during at least one of a "plan entry", "plan target", "navigate", "refine" or "deliver therapy" step.
47. A system according to Claim 31, wherein the trajectory guide comprises a targeting canula with a central passage and at least one fiducial marker, and wherein the circuit is configured to request and/or obtain high resolution image data from the MR scanner of a slab that includes the at least one fiducial marker based on predefined physical characteristics of the targeting canula to determine an actual trajectory of the targeting canula.
48. A system according to Claim 47, wherein the at least one marker comprises a proximal marker and a distal marker, and wherein the circuit is configured to selectively request a slab of a single one of the proximal or distal markers at a single point in time.
49. A system according to Claim 31, wherein the trajectory guide comprises a targeting canula with a central passage and spaced apart proximal and distal fiducial markers, the proximal marker having a different shape than the distal marker, and wherein the circuit is configured to request and/or obtain high resolution image data from the MR scanner of a slab that includes only one of the proximal or distal markers based on predefined physical characteristics of the targeting canula.
Applications Claiming Priority (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US97482107P | 2007-09-24 | 2007-09-24 | |
US60/974,821 | 2007-09-24 | ||
US12/134,412 US8175677B2 (en) | 2007-06-07 | 2008-06-06 | MRI-guided medical interventional systems and methods |
US12/134,412 | 2008-06-06 | ||
PCT/US2008/011050 WO2009042135A2 (en) | 2007-09-24 | 2008-09-24 | Mri surgical systems for real-time visualizations using mri image data and predefined data of surgical tools |
Publications (1)
Publication Number | Publication Date |
---|---|
CA2700577A1 true CA2700577A1 (en) | 2009-04-02 |
Family
ID=40515033
Family Applications (4)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CA2704739A Abandoned CA2704739A1 (en) | 2007-09-24 | 2008-09-24 | Control unit for mri-guided medical interventional systems |
CA2700529A Abandoned CA2700529A1 (en) | 2007-09-24 | 2008-09-24 | External mri imaging coil arrays and mri-guided interventional systems utilizing same |
CA2700577A Abandoned CA2700577A1 (en) | 2007-09-24 | 2008-09-24 | Mri surgical systems for real-time visualizations using mri image data and predefined data of surgical tools |
CA2700607A Abandoned CA2700607A1 (en) | 2007-09-24 | 2008-09-24 | Mri-compatible head fixation frame with cooperating head coil apparatus |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CA2704739A Abandoned CA2704739A1 (en) | 2007-09-24 | 2008-09-24 | Control unit for mri-guided medical interventional systems |
CA2700529A Abandoned CA2700529A1 (en) | 2007-09-24 | 2008-09-24 | External mri imaging coil arrays and mri-guided interventional systems utilizing same |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CA2700607A Abandoned CA2700607A1 (en) | 2007-09-24 | 2008-09-24 | Mri-compatible head fixation frame with cooperating head coil apparatus |
Country Status (6)
Country | Link |
---|---|
US (4) | US8175677B2 (en) |
EP (4) | EP2194894A1 (en) |
JP (1) | JP5632286B2 (en) |
CN (1) | CN101918855B (en) |
CA (4) | CA2704739A1 (en) |
WO (4) | WO2009042131A1 (en) |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8979871B2 (en) | 2009-08-13 | 2015-03-17 | Monteris Medical Corporation | Image-guided therapy of a tissue |
US9333038B2 (en) | 2000-06-15 | 2016-05-10 | Monteris Medical Corporation | Hyperthermia treatment and probe therefore |
US9433383B2 (en) | 2014-03-18 | 2016-09-06 | Monteris Medical Corporation | Image-guided therapy of a tissue |
US9504484B2 (en) | 2014-03-18 | 2016-11-29 | Monteris Medical Corporation | Image-guided therapy of a tissue |
US10327830B2 (en) | 2015-04-01 | 2019-06-25 | Monteris Medical Corporation | Cryotherapy, thermal therapy, temperature modulation therapy, and probe apparatus therefor |
US10675113B2 (en) | 2014-03-18 | 2020-06-09 | Monteris Medical Corporation | Automated therapy of a three-dimensional tissue region |
US11298043B2 (en) | 2016-08-30 | 2022-04-12 | The Regents Of The University Of California | Methods for biomedical targeting and delivery and devices and systems for practicing the same |
US11497576B2 (en) | 2017-07-17 | 2022-11-15 | Voyager Therapeutics, Inc. | Trajectory array guide system |
Families Citing this family (181)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040162637A1 (en) | 2002-07-25 | 2004-08-19 | Yulun Wang | Medical tele-robotic system with a master remote station with an arbitrator |
US7813836B2 (en) | 2003-12-09 | 2010-10-12 | Intouch Technologies, Inc. | Protocol for a remotely controlled videoconferencing robot |
US8077963B2 (en) | 2004-07-13 | 2011-12-13 | Yulun Wang | Mobile robot with a head-based movement mapping scheme |
US9198728B2 (en) | 2005-09-30 | 2015-12-01 | Intouch Technologies, Inc. | Multi-camera mobile teleconferencing platform |
US8849679B2 (en) | 2006-06-15 | 2014-09-30 | Intouch Technologies, Inc. | Remote controlled robot system that provides medical images |
US9160783B2 (en) | 2007-05-09 | 2015-10-13 | Intouch Technologies, Inc. | Robot system that operates through a network firewall |
US8175677B2 (en) * | 2007-06-07 | 2012-05-08 | MRI Interventions, Inc. | MRI-guided medical interventional systems and methods |
EP2174007A4 (en) * | 2007-07-02 | 2017-05-10 | Borgwarner Inc. | Inlet design for a pump assembly |
CA2700523A1 (en) * | 2007-09-24 | 2009-04-02 | Surgivision, Inc. | Mri-guided medical interventional systems and methods |
EP2192871B8 (en) | 2007-09-24 | 2015-01-28 | MRI Interventions, Inc. | Mri-compatible patch and method for identifying a position |
US8315689B2 (en) | 2007-09-24 | 2012-11-20 | MRI Interventions, Inc. | MRI surgical systems for real-time visualizations using MRI image data and predefined data of surgical tools |
US8340743B2 (en) * | 2007-11-21 | 2012-12-25 | MRI Interventions, Inc. | Methods, systems and computer program products for positioning a guidance apparatus relative to a patient |
US10875182B2 (en) | 2008-03-20 | 2020-12-29 | Teladoc Health, Inc. | Remote presence system mounted to operating room hardware |
US8179418B2 (en) | 2008-04-14 | 2012-05-15 | Intouch Technologies, Inc. | Robotic based health care system |
US8170241B2 (en) | 2008-04-17 | 2012-05-01 | Intouch Technologies, Inc. | Mobile tele-presence system with a microphone system |
US9193065B2 (en) | 2008-07-10 | 2015-11-24 | Intouch Technologies, Inc. | Docking system for a tele-presence robot |
US9842192B2 (en) | 2008-07-11 | 2017-12-12 | Intouch Technologies, Inc. | Tele-presence robot system with multi-cast features |
US8728092B2 (en) | 2008-08-14 | 2014-05-20 | Monteris Medical Corporation | Stereotactic drive system |
US8747418B2 (en) | 2008-08-15 | 2014-06-10 | Monteris Medical Corporation | Trajectory guide |
US20100063422A1 (en) * | 2008-09-08 | 2010-03-11 | Sunnybrook Health Sciences Center | Ultrasound therapy transducer head and ultrasound therapy system incorporating the same |
WO2010030373A2 (en) * | 2008-09-12 | 2010-03-18 | Surgivision, Inc. | Intrabody mri stacked flat loop antennas and related systems |
US8340819B2 (en) | 2008-09-18 | 2012-12-25 | Intouch Technologies, Inc. | Mobile videoconferencing robot system with network adaptive driving |
US8270698B2 (en) * | 2008-09-24 | 2012-09-18 | Merge Healthcare Incorporated | Anterior commissure and posterior commissure segmentation system and method |
US8996165B2 (en) | 2008-10-21 | 2015-03-31 | Intouch Technologies, Inc. | Telepresence robot with a camera boom |
US9138891B2 (en) | 2008-11-25 | 2015-09-22 | Intouch Technologies, Inc. | Server connectivity control for tele-presence robot |
US8463435B2 (en) | 2008-11-25 | 2013-06-11 | Intouch Technologies, Inc. | Server connectivity control for tele-presence robot |
US20100198052A1 (en) * | 2009-01-28 | 2010-08-05 | Kimble Jenkins | Mri-compatible articulating arms and related systems and methods |
US8849680B2 (en) | 2009-01-29 | 2014-09-30 | Intouch Technologies, Inc. | Documentation through a remote presence robot |
US20100217115A1 (en) * | 2009-02-25 | 2010-08-26 | Hushek Stephen G | Temperature sensing within a patient during mr imaging |
CN201384493Y (en) * | 2009-03-12 | 2010-01-20 | 西门子迈迪特(深圳)磁共振有限公司 | Positioning device used for magnetic resonance system |
US9232977B1 (en) * | 2009-03-27 | 2016-01-12 | Tausif-Ur Rehman | Instrument guiding device |
WO2010110881A1 (en) * | 2009-03-27 | 2010-09-30 | Hetherington Hoby P | Improved transceiver apparatus, system, and methodology for superior in-vivo imaging of human anatomy |
US8897920B2 (en) | 2009-04-17 | 2014-11-25 | Intouch Technologies, Inc. | Tele-presence robot system with software modularity, projector and laser pointer |
GB0908787D0 (en) | 2009-05-21 | 2009-07-01 | Renishaw Plc | Head clamp for imaging and neurosurgery |
GB0908784D0 (en) * | 2009-05-21 | 2009-07-01 | Renishaw Plc | Apparatus for imaging a body part |
US9259290B2 (en) | 2009-06-08 | 2016-02-16 | MRI Interventions, Inc. | MRI-guided surgical systems with proximity alerts |
JP2012529977A (en) | 2009-06-16 | 2012-11-29 | エムアールアイ・インターヴェンションズ,インコーポレイテッド | MRI guidance device and MRI guidance intervention system capable of tracking the device in near real time and generating a dynamic visualization of the device |
WO2010148095A2 (en) * | 2009-06-16 | 2010-12-23 | Neocoil, Llc | Modular apparatus for magnetic resonance imaging |
DE102009034312A1 (en) * | 2009-07-23 | 2011-01-27 | Martin Hempel | Protective housing for flexible MR coils |
WO2011013011A2 (en) * | 2009-07-29 | 2011-02-03 | Mauna Kea Technologies | Apparatus and method for brain fiber bundle microscopy |
US8384755B2 (en) | 2009-08-26 | 2013-02-26 | Intouch Technologies, Inc. | Portable remote presence robot |
US11399153B2 (en) | 2009-08-26 | 2022-07-26 | Teladoc Health, Inc. | Portable telepresence apparatus |
US9844414B2 (en) * | 2009-08-31 | 2017-12-19 | Gregory S. Fischer | System and method for robotic surgical intervention in a magnetic resonance imager |
US20110098553A1 (en) * | 2009-10-28 | 2011-04-28 | Steven Robbins | Automatic registration of images for image guided surgery |
US9864032B2 (en) * | 2010-01-05 | 2018-01-09 | National Health Research Institutes | Magnetic resonance imaging system |
US11154981B2 (en) | 2010-02-04 | 2021-10-26 | Teladoc Health, Inc. | Robot user interface for telepresence robot system |
US20110187875A1 (en) * | 2010-02-04 | 2011-08-04 | Intouch Technologies, Inc. | Robot face used in a sterile environment |
US8670017B2 (en) | 2010-03-04 | 2014-03-11 | Intouch Technologies, Inc. | Remote presence system including a cart that supports a robot face and an overhead camera |
CN102859385B (en) * | 2010-04-14 | 2015-06-17 | 皇家飞利浦电子股份有限公司 | Instrument guiding during magnetic resonance imaging |
EP2558154B1 (en) | 2010-04-16 | 2020-06-17 | ClearPoint Neuro, Inc. | Mri surgical systems including mri-compatible surgical cannulae for transferring a substance to and/or from a patient |
US9814885B2 (en) | 2010-04-27 | 2017-11-14 | Medtronic, Inc. | Stimulation electrode selection |
DE102010020152B4 (en) * | 2010-05-11 | 2013-10-24 | Siemens Aktiengesellschaft | Device with local coil arrangement and implantable device |
US10343283B2 (en) | 2010-05-24 | 2019-07-09 | Intouch Technologies, Inc. | Telepresence robot system that can be accessed by a cellular phone |
US10808882B2 (en) | 2010-05-26 | 2020-10-20 | Intouch Technologies, Inc. | Tele-robotic system with a robot face placed on a chair |
DE102010023844A1 (en) * | 2010-06-15 | 2011-12-15 | Siemens Aktiengesellschaft | MR RF coils with modulable flexibility |
EP2583244B1 (en) * | 2010-06-16 | 2019-07-24 | A² Surgical | Method of determination of access areas from 3d patient images |
US9264664B2 (en) | 2010-12-03 | 2016-02-16 | Intouch Technologies, Inc. | Systems and methods for dynamic bandwidth allocation |
KR102018763B1 (en) | 2011-01-28 | 2019-09-05 | 인터치 테크놀로지스 인코퍼레이티드 | Interfacing with a mobile telepresence robot |
US9323250B2 (en) | 2011-01-28 | 2016-04-26 | Intouch Technologies, Inc. | Time-dependent navigation of telepresence robots |
EP2675353A2 (en) * | 2011-02-17 | 2013-12-25 | MRI Interventions, Inc. | Thin-sleeve apparatus for reducing rf coupling of devices in mri environments |
KR101875122B1 (en) * | 2011-02-25 | 2018-07-11 | 주식회사 칼라세븐 | A Light therapy system using sunlight |
EP2494936A3 (en) * | 2011-03-03 | 2012-09-19 | Imris Inc. | MR compatible optical viewing device for use in the bore of an MR magnet |
WO2012127345A1 (en) * | 2011-03-18 | 2012-09-27 | Koninklijke Philips Electronics N.V. | Tracking brain deformation during neurosurgery |
CN104254296A (en) * | 2011-04-08 | 2014-12-31 | 曼特瑞斯医药有限责任公司 | Head fixation system and method |
US8406890B2 (en) * | 2011-04-14 | 2013-03-26 | Medtronic, Inc. | Implantable medical devices storing graphics processing data |
US9615770B2 (en) | 2011-04-15 | 2017-04-11 | Neocoil, Llc | Pediatric imaging assembly |
US10769739B2 (en) | 2011-04-25 | 2020-09-08 | Intouch Technologies, Inc. | Systems and methods for management of information among medical providers and facilities |
US20140139616A1 (en) | 2012-01-27 | 2014-05-22 | Intouch Technologies, Inc. | Enhanced Diagnostics for a Telepresence Robot |
US9098611B2 (en) | 2012-11-26 | 2015-08-04 | Intouch Technologies, Inc. | Enhanced video interaction for a user interface of a telepresence network |
DE102011079565B4 (en) * | 2011-07-21 | 2022-09-15 | Siemens Healthcare Gmbh | Direct connection head coil with height adjustment for e.g. Bechterew patients |
TWI442905B (en) * | 2011-09-30 | 2014-07-01 | Univ Nat Chiao Tung | Apparatus for locating the target of the stimulation |
US8836751B2 (en) | 2011-11-08 | 2014-09-16 | Intouch Technologies, Inc. | Tele-presence system with a user interface that displays different communication links |
JP6220345B2 (en) | 2011-12-05 | 2017-10-25 | コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. | Surgical tool positioning and orientation during patient-specific port placement |
RU2634296C2 (en) * | 2012-01-03 | 2017-10-24 | Конинклейке Филипс Н.В. | Device for position determination |
TWI463964B (en) * | 2012-03-03 | 2014-12-11 | Univ China Medical | System and apparatus for an image guided navigation system in surgery |
US8902278B2 (en) | 2012-04-11 | 2014-12-02 | Intouch Technologies, Inc. | Systems and methods for visualizing and managing telepresence devices in healthcare networks |
US9251313B2 (en) | 2012-04-11 | 2016-02-02 | Intouch Technologies, Inc. | Systems and methods for visualizing and managing telepresence devices in healthcare networks |
EP2666428B1 (en) * | 2012-05-21 | 2015-10-28 | Universität Bern | System and method for estimating the spatial position of a tool within an object |
WO2013176758A1 (en) | 2012-05-22 | 2013-11-28 | Intouch Technologies, Inc. | Clinical workflows utilizing autonomous and semi-autonomous telemedicine devices |
US9361021B2 (en) | 2012-05-22 | 2016-06-07 | Irobot Corporation | Graphical user interfaces including touchpad driving interfaces for telemedicine devices |
US9192393B2 (en) | 2012-05-31 | 2015-11-24 | MRI Interventions, Inc. | MRI compatible surgical drills and related methods |
US9498290B2 (en) | 2012-07-19 | 2016-11-22 | MRI Interventions, Inc. | Surgical navigation devices and methods |
US9610048B2 (en) | 2012-08-09 | 2017-04-04 | MRI Interventions, Inc. | Fiber optic systems for MRI suites and related devices and methods |
US10136955B2 (en) * | 2012-08-24 | 2018-11-27 | University Of Houston System | Robotic device for image-guided surgery and interventions |
MX2015002400A (en) * | 2012-08-24 | 2015-11-09 | Univ Houston | Robotic device and systems for image-guided and robot-assisted surgery. |
US9192446B2 (en) | 2012-09-05 | 2015-11-24 | MRI Interventions, Inc. | Trajectory guide frame for MRI-guided surgeries |
CN203101610U (en) * | 2012-09-25 | 2013-07-31 | 西门子(深圳)磁共振有限公司 | Receiving coil of magnetic resonance system |
US10588597B2 (en) * | 2012-12-31 | 2020-03-17 | Intuitive Surgical Operations, Inc. | Systems and methods for interventional procedure planning |
US9078588B2 (en) | 2013-03-08 | 2015-07-14 | MRI Interventions, Inc. | MRI compatible intrabody stylets and related methods and systems |
US9222996B2 (en) * | 2013-03-15 | 2015-12-29 | The Brigham And Women's Hospital, Inc. | Needle placement manipulator with two rotary guides |
EP2968857B1 (en) * | 2013-03-15 | 2022-05-04 | Intuitive Surgical Operations, Inc. | Shape sensor systems for tracking interventional instruments |
US10274553B2 (en) * | 2013-03-15 | 2019-04-30 | Canon U.S.A., Inc. | Needle placement manipulator with attachment for RF-coil |
US9730762B2 (en) | 2013-03-15 | 2017-08-15 | Neocoil, Llc | Automatic needle insertion location identification |
WO2014139024A1 (en) | 2013-03-15 | 2014-09-18 | Synaptive Medical (Barbados) Inc. | Planning, navigation and simulation systems and methods for minimally invasive therapy |
US9782198B2 (en) * | 2013-03-28 | 2017-10-10 | Koninklijke Philips N.V. | Localization of robotic remote center of motion point using custom trocar |
CN103340685A (en) * | 2013-06-27 | 2013-10-09 | 苏州边枫电子科技有限公司 | Auxiliary needle feeding device of pneumatic type puncture needle |
US9600778B2 (en) | 2013-07-02 | 2017-03-21 | Surgical Information Sciences, Inc. | Method for a brain region location and shape prediction |
WO2015010189A1 (en) | 2013-07-24 | 2015-01-29 | Centre For Surgical Invention & Innovation | Multi-function mounting interface for an image-guided robotic system and quick release interventional toolset |
US9891296B2 (en) | 2013-09-13 | 2018-02-13 | MRI Interventions, Inc. | Intrabody fluid transfer devices, systems and methods |
CA2924230C (en) * | 2013-09-18 | 2020-03-31 | iMIRGE Medical INC. | Optical targeting and visusalization of trajectories |
WO2015058815A1 (en) * | 2013-10-25 | 2015-04-30 | Brainlab Ag | Magnetic resonance coil unit and method for its manufacture |
CN103646182B (en) * | 2013-12-13 | 2018-11-02 | 陕西理工学院 | A kind of application process of flexibility dynamic and visual technology on the medical image |
US10595744B2 (en) | 2014-02-14 | 2020-03-24 | MRI Interventions, Inc. | Surgical tool-positioning devices and related methods |
US9867667B2 (en) | 2014-02-27 | 2018-01-16 | Canon Usa Inc. | Placement apparatus |
WO2015171988A1 (en) | 2014-05-09 | 2015-11-12 | Canon U.S.A., Inc. | Positioning apparatus |
MX366786B (en) | 2014-09-05 | 2019-07-23 | Hyperfine Res Inc | Noise suppression methods and apparatus. |
EP3190996A4 (en) | 2014-09-12 | 2018-04-25 | Canon U.S.A., Inc. | Needle positioning apparatus |
US10866291B2 (en) * | 2014-09-12 | 2020-12-15 | Emory University | Devices and systems for MRI-guided procedures |
US10813564B2 (en) | 2014-11-11 | 2020-10-27 | Hyperfine Research, Inc. | Low field magnetic resonance methods and apparatus |
EP3632923A1 (en) | 2015-01-16 | 2020-04-08 | Voyager Therapeutics, Inc. | Central nervous system targeting polynucleotides |
US10542961B2 (en) | 2015-06-15 | 2020-01-28 | The Research Foundation For The State University Of New York | System and method for infrasonic cardiac monitoring |
US10420626B2 (en) | 2015-06-30 | 2019-09-24 | Canon U.S.A., Inc. | Fiducial markers, systems, and methods of registration |
EP3316784B1 (en) * | 2015-06-30 | 2021-08-04 | Canon U.S.A., Inc. | Fiducial markers, systems, and methods of registration |
US9867673B2 (en) | 2015-07-14 | 2018-01-16 | Canon U.S.A, Inc. | Medical support device |
US10639065B2 (en) | 2015-07-21 | 2020-05-05 | Canon U.S.A., Inc. | Medical assist device |
US10324594B2 (en) * | 2015-10-30 | 2019-06-18 | Siemens Healthcare Gmbh | Enterprise protocol management |
CN108348305A (en) * | 2015-11-16 | 2018-07-31 | 思想外科有限公司 | Method for confirming the registration for being tracked bone |
JP6899387B2 (en) * | 2015-11-30 | 2021-07-07 | コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. | Clinical discovery wheel, system for searching clinical concepts |
JP6906239B2 (en) * | 2015-12-28 | 2021-07-21 | ザクト ロボティクス リミテッド | Adjustable registration frame |
USD824027S1 (en) | 2016-01-13 | 2018-07-24 | MRI Interventions, Inc. | Fins for a support column for a surgical trajectory frame |
USD829904S1 (en) | 2016-01-13 | 2018-10-02 | MRI Interventions, Inc. | Curved bracket for surgical navigation systems |
US10376333B2 (en) | 2016-01-14 | 2019-08-13 | MRI Interventions, Inc. | Devices for surgical navigation systems |
US10765489B2 (en) | 2016-01-29 | 2020-09-08 | Canon U.S.A., Inc. | Tool placement manipulator |
CN105726124B (en) * | 2016-02-06 | 2019-01-15 | 齐欣 | Near end of thighbone location of operation orientation system and preparation method thereof |
WO2017142698A1 (en) | 2016-02-17 | 2017-08-24 | MRI Interventions, Inc. | Intrabody surgical fluid transfer assemblies with adjustable exposed cannula to needle tip length, related systems and methods |
CN109106454B (en) * | 2016-05-31 | 2020-09-22 | 万伟东 | RC intracranial minimally invasive treatment positioning device |
CN106175893B (en) * | 2016-08-03 | 2018-10-30 | 福建医科大学附属第一医院 | A kind of device of arching trajectory implantation intracranial electrode |
CN106137283B (en) * | 2016-08-27 | 2018-08-17 | 天津大学 | A kind of natural cavity apparatus work rotary switch with regulatory function |
DE102016216203A1 (en) * | 2016-08-29 | 2017-09-14 | Siemens Healthcare Gmbh | Medical imaging system |
JP6948389B2 (en) | 2016-10-19 | 2021-10-13 | キヤノン ユーエスエイ, インコーポレイテッドCanon U.S.A., Inc | Placement manipulators and attachments for positioning puncture devices |
US10627464B2 (en) | 2016-11-22 | 2020-04-21 | Hyperfine Research, Inc. | Low-field magnetic resonance imaging methods and apparatus |
US10539637B2 (en) | 2016-11-22 | 2020-01-21 | Hyperfine Research, Inc. | Portable magnetic resonance imaging methods and apparatus |
US11842030B2 (en) | 2017-01-31 | 2023-12-12 | Medtronic Navigation, Inc. | Method and apparatus for image-based navigation |
CN106618507A (en) * | 2017-03-13 | 2017-05-10 | 北京理工大学 | Touch cerebration stimulation device and fine adjustment assembly |
EP3595568A1 (en) * | 2017-03-15 | 2020-01-22 | Orthotaxy | System for guiding a surgical tool relative to a target axis in spine surgery |
EP3613057A4 (en) * | 2017-04-18 | 2021-04-21 | Intuitive Surgical Operations, Inc. | Graphical user interface for planning a procedure |
US11862302B2 (en) | 2017-04-24 | 2024-01-02 | Teladoc Health, Inc. | Automated transcription and documentation of tele-health encounters |
US11690645B2 (en) | 2017-05-03 | 2023-07-04 | Medtronic Vascular, Inc. | Tissue-removing catheter |
CN114948106A (en) | 2017-05-03 | 2022-08-30 | 美敦力瓦斯科尔勒公司 | Tissue removal catheter with guidewire isolation bushing |
US10548674B2 (en) * | 2017-07-06 | 2020-02-04 | YellowDot Innovations, LLC | Robotic guide for medical device |
US10483007B2 (en) | 2017-07-25 | 2019-11-19 | Intouch Technologies, Inc. | Modular telehealth cart with thermal imaging and touch screen user interface |
EP3808849A1 (en) | 2017-08-03 | 2021-04-21 | Voyager Therapeutics, Inc. | Compositions and methods for delivery of aav |
US11202652B2 (en) | 2017-08-11 | 2021-12-21 | Canon U.S.A., Inc. | Registration and motion compensation for patient-mounted needle guide |
US10987016B2 (en) * | 2017-08-23 | 2021-04-27 | The Boeing Company | Visualization system for deep brain stimulation |
US11636944B2 (en) | 2017-08-25 | 2023-04-25 | Teladoc Health, Inc. | Connectivity infrastructure for a telehealth platform |
CN109620407B (en) * | 2017-10-06 | 2024-02-06 | 皇家飞利浦有限公司 | Treatment trajectory guidance system |
US10893911B2 (en) | 2017-11-26 | 2021-01-19 | Canon U.S.A., Inc. | Automated image cropping for enhanced automatic device-to-image registration |
WO2019168973A1 (en) * | 2018-02-27 | 2019-09-06 | The Regents Of The University Of Colorado, A Body Corporate | Robotic stereotaxic platform with computer vision |
FR3078879B1 (en) * | 2018-03-14 | 2020-03-06 | Assistance Publique Hopitaux De Paris | SURGICAL KIT TO BE USED DURING A CRANIECTOMY PROCEDURE |
US10617299B2 (en) | 2018-04-27 | 2020-04-14 | Intouch Technologies, Inc. | Telehealth cart that supports a removable tablet with seamless audio/video switching |
EP3781074A1 (en) | 2018-05-09 | 2021-02-24 | ClearPoint Neuro, Inc. | Mri compatible intrabody fluid transfer systems and related devices and methods |
US11253237B2 (en) | 2018-05-09 | 2022-02-22 | Clearpoint Neuro, Inc. | MRI compatible intrabody fluid transfer systems and related devices and methods |
TW202015742A (en) | 2018-05-15 | 2020-05-01 | 美商航海家醫療公司 | Compositions and methods for delivery of aav |
CA3099306A1 (en) | 2018-05-15 | 2019-11-21 | Voyager Therapeutics, Inc. | Compositions and methods for the treatment of parkinson's disease |
EP3793615A2 (en) | 2018-05-16 | 2021-03-24 | Voyager Therapeutics, Inc. | Directed evolution of aav to improve tropism for cns |
US11209509B2 (en) * | 2018-05-16 | 2021-12-28 | Viewray Technologies, Inc. | Resistive electromagnet systems and methods |
CN108814658B (en) * | 2018-06-27 | 2021-10-01 | 中国人民解放军陆军军医大学第三附属医院(野战外科研究所) | Craniotomy device and craniotomy method |
EP3826719A1 (en) | 2018-07-24 | 2021-06-02 | Voyager Therapeutics, Inc. | Systems and methods for producing gene therapy formulations |
WO2020077165A1 (en) | 2018-10-12 | 2020-04-16 | Voyager Therapeutics, Inc. | Compositions and methods for delivery of aav |
WO2020102729A1 (en) | 2018-11-16 | 2020-05-22 | Medtronic Vascular, Inc. | Tissue-removing catheter |
US11744655B2 (en) | 2018-12-04 | 2023-09-05 | Globus Medical, Inc. | Drill guide fixtures, cranial insertion fixtures, and related methods and robotic systems |
US11364086B2 (en) | 2019-02-01 | 2022-06-21 | Advanced Neuromodulation Systems, Inc. | Trajectory guide with dual arc arrangement |
US11298204B2 (en) * | 2019-02-01 | 2022-04-12 | Advanced Neuromodulation Systems, Inc. | Trajectory guide with dual gimbal drive arrangement |
US11406470B2 (en) | 2019-02-01 | 2022-08-09 | Advanced Neuromodulation Systems, Inc. | Trajectory guide with double X-Y sliding tables |
SG11202109927YA (en) * | 2019-03-25 | 2021-10-28 | Promaxo Inc | Single-sided fast mri gradient field coils and applications thereof |
US11819236B2 (en) | 2019-05-17 | 2023-11-21 | Medtronic Vascular, Inc. | Tissue-removing catheter |
US20220333133A1 (en) | 2019-09-03 | 2022-10-20 | Voyager Therapeutics, Inc. | Vectorized editing of nucleic acids to correct overt mutations |
US11684750B2 (en) | 2019-10-08 | 2023-06-27 | Clearpoint Neuro, Inc. | Extension tube assembly and related medical fluid transfer systems and methods |
WO2021113370A1 (en) | 2019-12-02 | 2021-06-10 | The General Hospital Corporation | Systems and methods for multi-modal bioimaging data integration and visualization |
US20230056943A1 (en) * | 2019-12-13 | 2023-02-23 | Dinesh Vyas | Stapler apparatus and methods for use |
EP4041105A1 (en) | 2019-12-19 | 2022-08-17 | ClearPoint Neuro, Inc. | Front-loadable fluid transfer assemblies and related medical fluid transfer systems |
US11602400B2 (en) * | 2020-01-10 | 2023-03-14 | Stryker European Operations Limited | Surgical arm and method of providing visual guidance for operating same |
US11925511B2 (en) | 2020-01-31 | 2024-03-12 | Clearpoint Neuro, Inc. | Surgical tool support systems including elongate support legs with adjustable lengths and related methods |
US20210282866A1 (en) * | 2020-03-12 | 2021-09-16 | Clearpoint Neuro, Inc. | Image-guided surgical systems with automated trajectory guide systems and related devices and methods |
CN113940754B (en) * | 2020-07-15 | 2024-03-22 | 台北医学大学 | Medical image processing system and method |
TWI737404B (en) * | 2020-07-15 | 2021-08-21 | 臺北醫學大學 | Medical image processing system and method thereof |
US11238591B1 (en) | 2020-07-15 | 2022-02-01 | Taipei Medical University (Tmu) | Medical image processing system and method thereof |
WO2022197380A1 (en) * | 2021-03-16 | 2022-09-22 | Clearpoint Neuro, Inc. | Directional-device intrabody placement systems and related methods |
EP4329642A1 (en) * | 2021-03-31 | 2024-03-06 | Clear Guide Medical, Inc. | System and method for image guided interventions |
CN113764074B (en) * | 2021-09-13 | 2024-01-02 | 杭州太美星程医药科技有限公司 | Image processing method and device, computer equipment and storage medium |
DE102022125703A1 (en) * | 2022-10-05 | 2024-04-11 | Otto-von-Guericke-Universität Magdeburg, Körperschaft des öffentlichen Rechts | COUPLING STATION AND REMOTE MANIPULATION SYSTEM |
Family Cites Families (144)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US2697433A (en) | 1951-12-04 | 1954-12-21 | Max A Zehnder | Device for accurately positioning and guiding guide wires used in the nailing of thefemoral neck |
US3878830A (en) * | 1973-05-31 | 1975-04-22 | Mediscience Technology Corp | Catheter system for blood gas monitoring |
US4051845A (en) * | 1976-03-05 | 1977-10-04 | The Kendall Company | Drape assembly with pouch and method |
US4386602A (en) | 1977-05-17 | 1983-06-07 | Sheldon Charles H | Intracranial surgical operative apparatus |
US4209258A (en) * | 1978-02-14 | 1980-06-24 | Oakes W Peter | Automatic continuous mixer apparatus |
US4319136A (en) | 1979-11-09 | 1982-03-09 | Jinkins J Randolph | Computerized tomography radiograph data transfer cap |
US4276697A (en) * | 1980-04-21 | 1981-07-07 | Astek Engineering, Inc. | Compliance element for remote center compliance unit |
US4838265A (en) | 1985-05-24 | 1989-06-13 | Cosman Eric R | Localization device for probe placement under CT scanner imaging |
US5085219A (en) * | 1987-10-30 | 1992-02-04 | The Regents Of The University Of California | Adjustable holders for magnetic reasonance imaging rf surface coil |
DE3804491A1 (en) * | 1987-12-02 | 1989-06-15 | Olympus Optical Co | Device for brain surgery |
JPH0252642A (en) | 1988-08-17 | 1990-02-22 | Olympus Optical Co Ltd | Brain operation equipment |
JPH0298335A (en) * | 1988-10-04 | 1990-04-10 | Toshiba Corp | Treatment portion positioning mechanism for magnetic resonance imaging |
US5125888A (en) | 1990-01-10 | 1992-06-30 | University Of Virginia Alumni Patents Foundation | Magnetic stereotactic system for treatment delivery |
US6006126A (en) | 1991-01-28 | 1999-12-21 | Cosman; Eric R. | System and method for stereotactic registration of image scan data |
JPH0584309A (en) | 1991-09-27 | 1993-04-06 | Olympus Optical Co Ltd | Water supply connector device of medical instrument |
US5469847A (en) | 1992-09-09 | 1995-11-28 | Izi Corporation | Radiographic multi-modality skin markers |
US5342356A (en) | 1992-12-02 | 1994-08-30 | Ellman Alan G | Electrical coupling unit for electrosurgery |
US5799099A (en) * | 1993-02-12 | 1998-08-25 | George S. Allen | Automatic technique for localizing externally attached fiducial markers in volume images of the head |
US6419680B1 (en) | 1993-06-10 | 2002-07-16 | Sherwood Services Ag | CT and MRI visible index markers for stereotactic localization |
US5469353A (en) | 1993-11-26 | 1995-11-21 | Access Radiology Corp. | Radiological image interpretation apparatus and method |
US5450293A (en) * | 1993-12-30 | 1995-09-12 | Hoffman; Elliott S. | Finger mounted fiber optic illumination system |
US5537704A (en) * | 1994-07-19 | 1996-07-23 | Ohio Medical Instrument Company, Inc. | Radiolucent head clamp |
US5803089A (en) * | 1994-09-15 | 1998-09-08 | Visualization Technology, Inc. | Position tracking and imaging system for use in medical applications |
US5728079A (en) | 1994-09-19 | 1998-03-17 | Cordis Corporation | Catheter which is visible under MRI |
US5695501A (en) | 1994-09-30 | 1997-12-09 | Ohio Medical Instrument Company, Inc. | Apparatus for neurosurgical stereotactic procedures |
US5971997A (en) * | 1995-02-03 | 1999-10-26 | Radionics, Inc. | Intraoperative recalibration apparatus for stereotactic navigators |
US5971984A (en) | 1995-03-01 | 1999-10-26 | Smith & Nephew, Inc. | Method of using an orthopaedic fixation device |
US5699801A (en) | 1995-06-01 | 1997-12-23 | The Johns Hopkins University | Method of internal magnetic resonance imaging and spectroscopic analysis and associated apparatus |
US5697958A (en) | 1995-06-07 | 1997-12-16 | Intermedics, Inc. | Electromagnetic noise detector for implantable medical devices |
US5638819A (en) | 1995-08-29 | 1997-06-17 | Manwaring; Kim H. | Method and apparatus for guiding an instrument to a target |
US5855582A (en) | 1995-12-19 | 1999-01-05 | Gildenberg; Philip L. | Noninvasive stereotactic apparatus and method for relating data between medical devices |
US5800353A (en) * | 1996-02-12 | 1998-09-01 | Mclaurin, Jr.; Robert L. | Automatic image registration of magnetic resonance imaging scans for localization, 3-dimensional treatment planning, and radiation treatment of abnormal lesions |
US5735278A (en) | 1996-03-15 | 1998-04-07 | National Research Council Of Canada | Surgical procedure with magnetic resonance imaging |
US5928145A (en) | 1996-04-25 | 1999-07-27 | The Johns Hopkins University | Method of magnetic resonance imaging and spectroscopic analysis and associated apparatus employing a loopless antenna |
US6675033B1 (en) | 1999-04-15 | 2004-01-06 | Johns Hopkins University School Of Medicine | Magnetic resonance imaging guidewire probe |
US6263229B1 (en) | 1998-11-13 | 2001-07-17 | Johns Hopkins University School Of Medicine | Miniature magnetic resonance catheter coils and related methods |
DE19618945C2 (en) | 1996-05-10 | 2003-02-27 | Phonak Ag Staefa | Fixable positioning system for a firm, play-free connection to the human skull |
DE19625834A1 (en) * | 1996-06-27 | 1998-01-02 | Siemens Ag | Medical system architecture |
US6177797B1 (en) | 1996-12-19 | 2001-01-23 | Advanced Imaging Research, Inc. | Radio-frequency coil and method for resonance/imaging analysis |
US5961455A (en) | 1996-12-31 | 1999-10-05 | Daum Gmbh | Device for positioning a medical instrument and method |
US6752812B1 (en) | 1997-05-15 | 2004-06-22 | Regent Of The University Of Minnesota | Remote actuation of trajectory guide |
US6267769B1 (en) | 1997-05-15 | 2001-07-31 | Regents Of The Universitiy Of Minnesota | Trajectory guide method and apparatus for use in magnetic resonance and computerized tomographic scanners |
JP2001525703A (en) | 1997-05-15 | 2001-12-11 | リージェンツ オブ ザ ユニバーシティ オブ ミネソタ | Method and apparatus for targeted dosing of patients using magnetic resonance imaging |
US5993463A (en) | 1997-05-15 | 1999-11-30 | Regents Of The University Of Minnesota | Remote actuation of trajectory guide |
US6050992A (en) | 1997-05-19 | 2000-04-18 | Radiotherapeutics Corporation | Apparatus and method for treating tissue with multiple electrodes |
SE9800126D0 (en) | 1998-01-20 | 1998-01-20 | Pacesetter Ab | Implantable medical device |
US6360116B1 (en) | 1998-02-27 | 2002-03-19 | Varian Medical Systems, Inc. | Brachytherapy system for prostate cancer treatment with computer implemented systems and processes to facilitate pre-operative planning and post-operative evaluations |
US6273896B1 (en) | 1998-04-21 | 2001-08-14 | Neutar, Llc | Removable frames for stereotactic localization |
US6298262B1 (en) | 1998-04-21 | 2001-10-02 | Neutar, Llc | Instrument guidance for stereotactic surgery |
US6529765B1 (en) | 1998-04-21 | 2003-03-04 | Neutar L.L.C. | Instrumented and actuated guidance fixture for sterotactic surgery |
US6546277B1 (en) * | 1998-04-21 | 2003-04-08 | Neutar L.L.C. | Instrument guidance system for spinal and other surgery |
FR2779339B1 (en) | 1998-06-09 | 2000-10-13 | Integrated Surgical Systems Sa | MATCHING METHOD AND APPARATUS FOR ROBOTIC SURGERY, AND MATCHING DEVICE COMPRISING APPLICATION |
US6459927B1 (en) | 1999-07-06 | 2002-10-01 | Neutar, Llc | Customizable fixture for patient positioning |
US6351662B1 (en) | 1998-08-12 | 2002-02-26 | Neutar L.L.C. | Movable arm locator for stereotactic surgery |
US6282437B1 (en) | 1998-08-12 | 2001-08-28 | Neutar, Llc | Body-mounted sensing system for stereotactic surgery |
US6117143A (en) | 1998-09-11 | 2000-09-12 | Hybex Surgical Specialties, Inc. | Apparatus for frameless stereotactic surgery |
US20050240445A1 (en) | 1998-09-29 | 2005-10-27 | Michael Sutherland | Medical archive library and method |
US6954802B2 (en) * | 1998-09-29 | 2005-10-11 | Tdk Electronics Corporation | Removable media recording station for the medical industry |
US6195577B1 (en) | 1998-10-08 | 2001-02-27 | Regents Of The University Of Minnesota | Method and apparatus for positioning a device in a body |
US7844319B2 (en) | 1998-11-04 | 2010-11-30 | Susil Robert C | Systems and methods for magnetic-resonance-guided interventional procedures |
US6701176B1 (en) | 1998-11-04 | 2004-03-02 | Johns Hopkins University School Of Medicine | Magnetic-resonance-guided imaging, electrophysiology, and ablation |
US6198961B1 (en) * | 1998-11-12 | 2001-03-06 | Picker International, Inc. | Interventional radio frequency coil assembly for magnetic resonance (MR) guided neurosurgery |
WO2000028900A1 (en) * | 1998-11-17 | 2000-05-25 | Metra Biosystems, Inc. | Test object geometry for ultrasound transmission calibration |
US6798206B2 (en) | 1998-11-25 | 2004-09-28 | Medrad, Inc. | Neurovascular coil system and interface and system therefor and method of operating same in a multitude of modes |
US6198962B1 (en) * | 1998-11-25 | 2001-03-06 | Toshiba America Mri, Inc. | Quadrature detection coil for interventional MRI |
US6284971B1 (en) | 1998-11-25 | 2001-09-04 | Johns Hopkins University School Of Medicine | Enhanced safety coaxial cables |
US6470207B1 (en) | 1999-03-23 | 2002-10-22 | Surgical Navigation Technologies, Inc. | Navigational guidance via computer-assisted fluoroscopic imaging |
US7848788B2 (en) | 1999-04-15 | 2010-12-07 | The Johns Hopkins University | Magnetic resonance imaging probe |
US6606513B2 (en) | 2000-02-01 | 2003-08-12 | Surgi-Vision, Inc. | Magnetic resonance imaging transseptal needle antenna |
US6491699B1 (en) | 1999-04-20 | 2002-12-10 | Surgical Navigation Technologies, Inc. | Instrument guidance method and system for image guided surgery |
US6539263B1 (en) | 1999-06-11 | 2003-03-25 | Cornell Research Foundation, Inc. | Feedback mechanism for deep brain stimulation |
US6167311A (en) | 1999-06-14 | 2000-12-26 | Electro Core Techniques, Llc | Method of treating psychological disorders by brain stimulation within the thalamus |
US20030205233A1 (en) * | 1999-12-02 | 2003-11-06 | A-Med Systems, Inc. | Surgical drape and panel assembly |
JP2001161660A (en) | 1999-12-10 | 2001-06-19 | Hitachi Medical Corp | Magnetic resonance imaging system |
DK200001852A (en) | 1999-12-14 | 2001-06-15 | Asahi Optical Co Ltd | Manipulation section for an endoscopic treatment instrument |
US6438423B1 (en) | 2000-01-20 | 2002-08-20 | Electrocore Technique, Llc | Method of treating complex regional pain syndromes by electrical stimulation of the sympathetic nerve chain |
US6356786B1 (en) | 2000-01-20 | 2002-03-12 | Electrocore Techniques, Llc | Method of treating palmar hyperhydrosis by electrical stimulation of the sympathetic nervous chain |
US6708064B2 (en) | 2000-02-24 | 2004-03-16 | Ali R. Rezai | Modulation of the brain to affect psychiatric disorders |
US6609030B1 (en) | 2000-02-24 | 2003-08-19 | Electrocore Techniques, Llc | Method of treating psychiatric diseases by neuromodulation within the dorsomedial thalamus |
WO2001064124A1 (en) | 2000-03-01 | 2001-09-07 | Surgical Navigation Technologies, Inc. | Multiple cannula image guided tool for image guided procedures |
AU2001247806A1 (en) | 2000-03-24 | 2001-10-08 | Surgi-Vision | Endoluminal mri probe |
US6477399B2 (en) * | 2000-03-29 | 2002-11-05 | Mcw Research Foundation, Inc. | Method for determining the reliability of fMRI parameters |
DE10015670A1 (en) * | 2000-03-29 | 2001-10-11 | Forschungszentrum Juelich Gmbh | Head holder has arrangement for holding head in one position and at least three markers. whose positions can be freely selected in non-symmetrical arrangement with respect to each other |
US6315783B1 (en) * | 2000-04-07 | 2001-11-13 | Odin Technologies, Ltd. | Surgical head support |
US7366561B2 (en) | 2000-04-07 | 2008-04-29 | Medtronic, Inc. | Robotic trajectory guide |
US7660621B2 (en) | 2000-04-07 | 2010-02-09 | Medtronic, Inc. | Medical device introducer |
US6575904B2 (en) * | 2000-05-09 | 2003-06-10 | Matsushita Electric Industrial Co., Ltd. | Biodata interfacing system |
US6526318B1 (en) | 2000-06-16 | 2003-02-25 | Mehdi M. Ansarinia | Stimulation method for the sphenopalatine ganglia, sphenopalatine nerve, or vidian nerve for treatment of medical conditions |
DE10029737B4 (en) | 2000-06-23 | 2006-01-19 | Mri Devices Daum Gmbh | Navigation of a medical instrument |
DE10029736B4 (en) | 2000-06-23 | 2004-05-06 | Mri Devices Daum Gmbh | Minimally invasive neurosurgery access |
AU2001285071A1 (en) | 2000-08-17 | 2002-02-25 | John David | Trajectory guide with instrument immobilizer |
US6712773B1 (en) | 2000-09-11 | 2004-03-30 | Tyco Healthcare Group Lp | Biopsy system |
US6405079B1 (en) | 2000-09-22 | 2002-06-11 | Mehdi M. Ansarinia | Stimulation method for the dural venous sinuses and adjacent dura for treatment of medical conditions |
US6577888B1 (en) | 2000-09-29 | 2003-06-10 | Usa Instruments, Inc. | Sliding-dome and split-top MRI radio frequency quadrature array coil system |
US6591128B1 (en) * | 2000-11-09 | 2003-07-08 | Koninklijke Philips Electronics, N.V. | MRI RF coil systems having detachable, relocatable, and or interchangeable sections and MRI imaging systems and methods employing the same |
US7646898B1 (en) | 2000-11-24 | 2010-01-12 | Kent Ridge Digital Labs | Methods and apparatus for processing medical images |
US20030009095A1 (en) * | 2001-05-21 | 2003-01-09 | Skarda James R. | Malleable elongated medical device |
FI110478B (en) * | 2001-05-29 | 2003-02-14 | Planmeca Oy | Method and apparatus for defining a beam |
US7532920B1 (en) * | 2001-05-31 | 2009-05-12 | Advanced Cardiovascular Systems, Inc. | Guidewire with optical fiber |
US20030055436A1 (en) | 2001-09-14 | 2003-03-20 | Wolfgang Daum | Navigation of a medical instrument |
US6772000B2 (en) | 2001-10-19 | 2004-08-03 | Scimed Life Systems, Inc. | Magnetic resonance imaging devices with a contrast medium for improved imaging |
US7020844B2 (en) * | 2001-11-21 | 2006-03-28 | General Electric Company | Method and apparatus for managing workflow in prescribing and processing medical images |
JP4430937B2 (en) * | 2001-12-07 | 2010-03-10 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | Medical viewing system and method for spatial enhancement of structures in noisy images |
US20040092810A1 (en) | 2002-02-14 | 2004-05-13 | Wolfgang Daum | Method and apparatus for MR-guided biopsy |
JP2005523741A (en) * | 2002-04-22 | 2005-08-11 | ザ ジョンズ ホプキンス ユニバーシティ | Device for inserting a medical instrument during a medical imaging process |
US6725092B2 (en) | 2002-04-25 | 2004-04-20 | Biophan Technologies, Inc. | Electromagnetic radiation immune medical assist device adapter |
US7022082B2 (en) | 2002-05-13 | 2006-04-04 | Sonek Jiri D | Needle guide systems and methods |
US7894877B2 (en) * | 2002-05-17 | 2011-02-22 | Case Western Reserve University | System and method for adjusting image parameters based on device tracking |
CA2487140C (en) | 2002-05-29 | 2011-09-20 | Surgi-Vision, Inc. | Magnetic resonance probes |
AU2003245758A1 (en) * | 2002-06-21 | 2004-01-06 | Cedara Software Corp. | Computer assisted system and method for minimal invasive hip, uni knee and total knee replacement |
US6978167B2 (en) * | 2002-07-01 | 2005-12-20 | Claron Technology Inc. | Video pose tracking system and method |
US7720522B2 (en) | 2003-02-25 | 2010-05-18 | Medtronic, Inc. | Fiducial marker devices, tools, and methods |
AU2003257309A1 (en) | 2002-08-13 | 2004-02-25 | Microbotics Corporation | Microsurgical robot system |
EP1581100A4 (en) | 2002-09-30 | 2009-01-21 | Stereotaxis Inc | A method and apparatus for improved surgical navigation employing electronic identification with automatically actuated flexible medical devices |
US20060173283A1 (en) * | 2002-11-27 | 2006-08-03 | Oskar Axelsson | Method of magnetic resonance imaging |
US7636596B2 (en) | 2002-12-20 | 2009-12-22 | Medtronic, Inc. | Organ access device and method |
EP1596716B1 (en) | 2003-01-24 | 2014-04-30 | The General Hospital Corporation | System and method for identifying tissue using low-coherence interferometry |
US7559935B2 (en) * | 2003-02-20 | 2009-07-14 | Medtronic, Inc. | Target depth locators for trajectory guide for introducing an instrument |
US7896889B2 (en) | 2003-02-20 | 2011-03-01 | Medtronic, Inc. | Trajectory guide with angled or patterned lumens or height adjustment |
US7203551B2 (en) | 2003-04-25 | 2007-04-10 | Medtronic, Inc. | Implantable lead-based sensor powered by piezoelectric transformer |
US7167760B2 (en) | 2003-04-28 | 2007-01-23 | Vanderbilt University | Apparatus and methods of optimal placement of deep brain stimulator |
CA2473963A1 (en) | 2003-07-14 | 2005-01-14 | Sunnybrook And Women's College Health Sciences Centre | Optical image-based position tracking for magnetic resonance imaging |
US7313430B2 (en) | 2003-08-28 | 2007-12-25 | Medtronic Navigation, Inc. | Method and apparatus for performing stereotactic surgery |
EP1673146B1 (en) * | 2003-09-30 | 2012-11-14 | Koninklijke Philips Electronics N.V. | Target tracking apparatus for radiation treatment planning and delivery |
JP2005137498A (en) | 2003-11-05 | 2005-06-02 | Sanko Sangyo Co Ltd | Holder |
US20050131522A1 (en) * | 2003-12-10 | 2005-06-16 | Stinson Jonathan S. | Medical devices and methods of making the same |
US8620406B2 (en) * | 2004-01-23 | 2013-12-31 | Boston Scientific Scimed, Inc. | Medical devices visible by magnetic resonance imaging |
US7174219B2 (en) | 2004-03-30 | 2007-02-06 | Medtronic, Inc. | Lead electrode for use in an MRI-safe implantable medical device |
ATE444712T1 (en) | 2004-05-21 | 2009-10-15 | Ethicon Endo Surgery Inc | MRI BIOPSY DEVICE WITH A DISPLAYABLE PENETRATION PART |
US6939165B1 (en) * | 2004-07-22 | 2005-09-06 | Hon Hai Precision Ind. Co., Ltd. | Cable connector assembly with cable holder |
EP1786320B1 (en) | 2004-07-27 | 2016-09-14 | MRI Interventions, Inc. | Mri systems having mri compatible universal delivery cannulas with cooperating mri antenna probes and related systems and methods |
JP4912304B2 (en) | 2004-08-09 | 2012-04-11 | ザ ジョンズ ホプキンス ユニヴァーシティ | Implantable MRI compatible stimulation lead and antenna and related systems |
WO2006081409A2 (en) | 2005-01-28 | 2006-08-03 | Massachusetts General Hospital | Guidance and insertion system |
CN101553165B (en) | 2005-05-04 | 2011-05-18 | 波士顿科学神经调制公司 | Electrical lead for an electronic device such as an implantable device |
US7869858B2 (en) * | 2005-05-12 | 2011-01-11 | General Electric Company | Patient table system and apparatus |
US8211057B2 (en) | 2005-09-12 | 2012-07-03 | Nemoto Kyorindo Co., Ltd. | Chemical liquid injection system |
EP1924198B1 (en) | 2005-09-13 | 2019-04-03 | Veran Medical Technologies, Inc. | Apparatus for image guided accuracy verification |
WO2007056458A2 (en) | 2005-11-07 | 2007-05-18 | Vanderbilt University | Adjustable universal surgical platform |
WO2007064739A2 (en) * | 2005-11-29 | 2007-06-07 | Surgi-Vision, Inc. | Mri-guided localization and/or lead placement systems, related methods, devices and computer program products |
DE102005062716A1 (en) * | 2005-12-28 | 2007-07-05 | Precisis Ag | Stereotaxis frame for patient, is formed in anatomical form of skull cross-section that is oval and non-circular, rectangular or square to facilitate access to head coil in magnetic resonance imaging |
EP2001367A4 (en) | 2006-03-14 | 2010-11-24 | Univ Johns Hopkins | Apparatus for insertion of a medical device within a body during a medical imaging process and devices and methods related thereto |
US7359226B2 (en) * | 2006-08-28 | 2008-04-15 | Qimonda Ag | Transistor, memory cell array and method for forming and operating a memory device |
US8175677B2 (en) | 2007-06-07 | 2012-05-08 | MRI Interventions, Inc. | MRI-guided medical interventional systems and methods |
CA2700523A1 (en) | 2007-09-24 | 2009-04-02 | Surgivision, Inc. | Mri-guided medical interventional systems and methods |
-
2008
- 2008-06-06 US US12/134,412 patent/US8175677B2/en active Active
- 2008-09-24 US US12/237,075 patent/US9097756B2/en active Active
- 2008-09-24 WO PCT/US2008/011043 patent/WO2009042131A1/en active Application Filing
- 2008-09-24 WO PCT/US2008/011051 patent/WO2009042136A1/en active Application Filing
- 2008-09-24 EP EP08832787A patent/EP2194894A1/en not_active Withdrawn
- 2008-09-24 CA CA2704739A patent/CA2704739A1/en not_active Abandoned
- 2008-09-24 CN CN2008801173424A patent/CN101918855B/en not_active Expired - Fee Related
- 2008-09-24 WO PCT/US2008/011050 patent/WO2009042135A2/en active Application Filing
- 2008-09-24 CA CA2700529A patent/CA2700529A1/en not_active Abandoned
- 2008-09-24 US US12/237,102 patent/US7602190B2/en not_active Expired - Fee Related
- 2008-09-24 US US12/237,033 patent/US8208993B2/en active Active
- 2008-09-24 EP EP08833057.6A patent/EP2195676B1/en active Active
- 2008-09-24 EP EP08832936A patent/EP2193384A1/en not_active Withdrawn
- 2008-09-24 JP JP2010525854A patent/JP5632286B2/en active Active
- 2008-09-24 WO PCT/US2008/011073 patent/WO2009042152A1/en active Application Filing
- 2008-09-24 CA CA2700577A patent/CA2700577A1/en not_active Abandoned
- 2008-09-24 EP EP08833037A patent/EP2192870A1/en not_active Withdrawn
- 2008-09-24 CA CA2700607A patent/CA2700607A1/en not_active Abandoned
Cited By (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9333038B2 (en) | 2000-06-15 | 2016-05-10 | Monteris Medical Corporation | Hyperthermia treatment and probe therefore |
US9387042B2 (en) | 2000-06-15 | 2016-07-12 | Monteris Medical Corporation | Hyperthermia treatment and probe therefor |
US9510909B2 (en) | 2009-08-13 | 2016-12-06 | Monteris Medical Corporation | Image-guide therapy of a tissue |
US9211157B2 (en) | 2009-08-13 | 2015-12-15 | Monteris Medical Corporation | Probe driver |
US9271794B2 (en) | 2009-08-13 | 2016-03-01 | Monteris Medical Corporation | Monitoring and noise masking of thermal therapy |
US8979871B2 (en) | 2009-08-13 | 2015-03-17 | Monteris Medical Corporation | Image-guided therapy of a tissue |
US10610317B2 (en) | 2009-08-13 | 2020-04-07 | Monteris Medical Corporation | Image-guided therapy of a tissue |
US10188462B2 (en) | 2009-08-13 | 2019-01-29 | Monteris Medical Corporation | Image-guided therapy of a tissue |
US10548678B2 (en) | 2012-06-27 | 2020-02-04 | Monteris Medical Corporation | Method and device for effecting thermal therapy of a tissue |
US9433383B2 (en) | 2014-03-18 | 2016-09-06 | Monteris Medical Corporation | Image-guided therapy of a tissue |
US9700342B2 (en) | 2014-03-18 | 2017-07-11 | Monteris Medical Corporation | Image-guided therapy of a tissue |
US10092367B2 (en) | 2014-03-18 | 2018-10-09 | Monteris Medical Corporation | Image-guided therapy of a tissue |
US9504484B2 (en) | 2014-03-18 | 2016-11-29 | Monteris Medical Corporation | Image-guided therapy of a tissue |
US10342632B2 (en) | 2014-03-18 | 2019-07-09 | Monteris Medical Corporation | Image-guided therapy of a tissue |
US9492121B2 (en) | 2014-03-18 | 2016-11-15 | Monteris Medical Corporation | Image-guided therapy of a tissue |
US9486170B2 (en) | 2014-03-18 | 2016-11-08 | Monteris Medical Corporation | Image-guided therapy of a tissue |
US10675113B2 (en) | 2014-03-18 | 2020-06-09 | Monteris Medical Corporation | Automated therapy of a three-dimensional tissue region |
US10327830B2 (en) | 2015-04-01 | 2019-06-25 | Monteris Medical Corporation | Cryotherapy, thermal therapy, temperature modulation therapy, and probe apparatus therefor |
US11672583B2 (en) | 2015-04-01 | 2023-06-13 | Monteris Medical Corporation | Cryotherapy, thermal therapy, temperature modulation therapy, and probe apparatus therefor |
US11298043B2 (en) | 2016-08-30 | 2022-04-12 | The Regents Of The University Of California | Methods for biomedical targeting and delivery and devices and systems for practicing the same |
US11298041B2 (en) | 2016-08-30 | 2022-04-12 | The Regents Of The University Of California | Methods for biomedical targeting and delivery and devices and systems for practicing the same |
US11497576B2 (en) | 2017-07-17 | 2022-11-15 | Voyager Therapeutics, Inc. | Trajectory array guide system |
Also Published As
Publication number | Publication date |
---|---|
EP2195676B1 (en) | 2016-12-14 |
WO2009042136A1 (en) | 2009-04-02 |
US20080306375A1 (en) | 2008-12-11 |
CN101918855A (en) | 2010-12-15 |
EP2192870A1 (en) | 2010-06-09 |
EP2194894A1 (en) | 2010-06-16 |
CA2704739A1 (en) | 2009-04-02 |
WO2009042135A2 (en) | 2009-04-02 |
WO2009042135A3 (en) | 2009-08-06 |
US20090082783A1 (en) | 2009-03-26 |
US8208993B2 (en) | 2012-06-26 |
CA2700529A1 (en) | 2009-04-02 |
US20090079431A1 (en) | 2009-03-26 |
CN101918855B (en) | 2013-07-17 |
US8175677B2 (en) | 2012-05-08 |
EP2193384A1 (en) | 2010-06-09 |
CA2700607A1 (en) | 2009-04-02 |
US20090112082A1 (en) | 2009-04-30 |
WO2009042131A1 (en) | 2009-04-02 |
EP2195676A2 (en) | 2010-06-16 |
JP5632286B2 (en) | 2014-11-26 |
US9097756B2 (en) | 2015-08-04 |
US7602190B2 (en) | 2009-10-13 |
WO2009042152A1 (en) | 2009-04-02 |
JP2010540021A (en) | 2010-12-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11317982B2 (en) | Image processing circuits for real-time visualizations using MRI image data and predefined data of surgical tools | |
EP2195676B1 (en) | Mri surgical systems for real-time visualizations using mri image data and predefined data of surgical tools | |
US11793933B2 (en) | MRI-compatible surgical cannulae for transferring a substance to and/or from a patient | |
CA2931877C (en) | Trajectory guidance alignment system and methods | |
US10413366B2 (en) | Trajectory guidance alignment system and methods | |
US9082215B2 (en) | Method of and system for overlaying NBS functional data on a live image of a brain | |
EP3046500B1 (en) | Surgical navigation system and related device | |
US11622699B2 (en) | Trajectory alignment system and methods | |
CA2927381C (en) | Trajectory alignment system and methods | |
EP3892214B1 (en) | Mri surgical systems including mri-compatible surgical cannulas for transferring a substance to and/or from a patient | |
US20210343397A1 (en) | Surgical planning systems that automatically assess different potential trajectory paths and identify candidate trajectories for surgical systems | |
AU2022311784A1 (en) | Augmented reality-driven guidance for interventional procedures |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
EEER | Examination request |
Effective date: 20130903 |
|
FZDE | Dead |
Effective date: 20160926 |