US9014423B2 - Method and system for catheter tracking in fluoroscopic images using adaptive discriminant learning and measurement fusion - Google Patents

Method and system for catheter tracking in fluoroscopic images using adaptive discriminant learning and measurement fusion Download PDF

Info

Publication number
US9014423B2
US9014423B2 US13/412,882 US201213412882A US9014423B2 US 9014423 B2 US9014423 B2 US 9014423B2 US 201213412882 A US201213412882 A US 201213412882A US 9014423 B2 US9014423 B2 US 9014423B2
Authority
US
United States
Prior art keywords
model
online
tracking
discriminant
adaptive
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US13/412,882
Other versions
US20120238866A1 (en
Inventor
Peng Wang
Yefeng Zheng
Matthias John
Jan Boese
Gareth Funka-Lea
Dorin Comaniciu
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Siemens Healthineers AG
Original Assignee
Siemens AG
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Siemens AG filed Critical Siemens AG
Priority to US13/412,882 priority Critical patent/US9014423B2/en
Priority to CN201210120135.2A priority patent/CN102697482B/en
Assigned to SIEMENS CORPORATION reassignment SIEMENS CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: WANG, PENG, COMANICIU, DORIN, FUNKA-LEA, GARETH, ZHENG, YEFENG
Assigned to SIEMENS AKTIENGESELLSCHAFT reassignment SIEMENS AKTIENGESELLSCHAFT ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BOESE, JAN, JOHN, MATTHIAS
Publication of US20120238866A1 publication Critical patent/US20120238866A1/en
Assigned to SIEMENS AKTIENGESELLSCHAFT reassignment SIEMENS AKTIENGESELLSCHAFT ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SIEMENS CORPORATION
Publication of US9014423B2 publication Critical patent/US9014423B2/en
Application granted granted Critical
Assigned to SIEMENS HEALTHCARE GMBH reassignment SIEMENS HEALTHCARE GMBH ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SIEMENS AKTIENGESELLSCHAFT
Assigned to Siemens Healthineers Ag reassignment Siemens Healthineers Ag ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SIEMENS HEALTHCARE GMBH
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus for radiation diagnosis, e.g. combined with radiation therapy equipment
    • A61B6/48Diagnostic techniques
    • A61B6/486Diagnostic techniques involving generating temporal series of image data
    • A61B6/487Diagnostic techniques involving generating temporal series of image data involving fluoroscopy
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus for radiation diagnosis, e.g. combined with radiation therapy equipment
    • A61B6/12Devices for detecting or locating foreign bodies
    • G06T7/2033
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus for radiation diagnosis, e.g. combined with radiation therapy equipment
    • A61B6/50Clinical applications
    • A61B6/503Clinical applications involving diagnosis of heart
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10064Fluorescence image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20076Probabilistic image processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30021Catheter; Guide wire
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30048Heart; Cardiac

Definitions

  • the present invention relates to medical imaging, and more particularly, to automatic tracking of catheter motion in a fluoroscopic image sequence.
  • Aortic valve disease affects a large number of people globally and is the most common type of valvular disease in developed countries. Implantation of a prosthetic aortic valve is often necessary to replace a severely damaged native valve.
  • TAVI minimally invasive transcatheter aortic valve implantation
  • Interventional surgeries, such as TAVI are typically performed under the guidance of real time fluoroscopic (x-ray) images.
  • x-ray real time fluoroscopic
  • the aorta will be visible and an annulus line can be identified in a 2D fluoroscopic image.
  • the aorta and annulus line will not be visible. Accordingly, it is desirable to track the motion of the aorta in non-contrast enhanced fluoroscopic images in order to reduce the exposure of patients to the contrast agent.
  • the present invention provides a method and system for automatically tracking an object, such as a catheter, in a sequence of fluoroscopic images.
  • Embodiments of the present invention utilize adaptive discriminant learning and measurement fusion for image-based catheter tracking.
  • Embodiments of the present invention can be used to track a pigtail catheter in trans-catheter aortic valve implantation (TAVI). Since the pigtail catheter has the same motion as the aorta, it is possible to track the motion of the aorta by tracking the pigtail catheter.
  • TAVI trans-catheter aortic valve implantation
  • an adaptive discriminant model is trained online based on a tracked object, such as a pigtail catheter tip, in at least one previous frame of a fluoroscopic image sequence.
  • the object is tracked in the current frame of the fluoroscopic image sequence based at least on the adaptive discriminant model trained online.
  • the object may be tracked in the current frame based on a fusion of three types of measurement models including the adaptive discriminant model trained online, an object detection model trained offline, and an online appearance model.
  • FIG. 1 illustrates an example of motion compensation for aortic valve implantation by tracking a pigtail catheter
  • FIG. 2 illustrates examples of pigtail catheters in different fluoroscopic image sequences
  • FIG. 3 illustrates a sequence of fluoroscopic images in which a pigtail catheter is occluded by another device
  • FIG. 4 illustrates an example of adaptive discriminant learning for detection of a pigtail catheter in a fluoroscopic image sequence
  • FIG. 5 illustrates a method for tracking an object in a sequence of medical images according to an embodiment of the present invention
  • FIG. 6 illustrates an algorithm for performing the adaptive discriminant learning according to an embodiment of the present invention
  • FIG. 7 illustrates exemplary pigtail catheter tracking results
  • FIG. 8 is a high level block diagram of a computer capable of implementing the present invention.
  • the present invention is directed to a method and system for automatically tracking an object, such as a catheter, in a fluoroscopic image sequence.
  • a digital image is often composed of digital representations of one or more objects (or shapes).
  • the digital representation of an object is often described herein in terms of identifying and manipulating the objects.
  • Such manipulations are virtual manipulations accomplished in the memory or other circuitry/hardware of a computer system. Accordingly, it is to be understood that embodiments of the present invention may be performed within a computer system using data stored within the computer system.
  • Embodiments of the present invention provide a computational framework for robustly tracking an object by adaptive discriminant learning and measurement fusion.
  • Such a tracking framework can be applied in Trans-catheter Aortic Valve Implantation (TAVI) in order to track a pigtail catheter.
  • a medical device such as a pigtail catheter can be tracked in a TAVI procedure to provide motion compensation for a 2D/3D overlay and reduce the exposure of a patient to contrast agent.
  • FIG. 1 illustrates an example of motion compensation for aortic valve implantation by tracking a pigtail catheter.
  • a pigtail catheter is a type of medical device that is inserted into the aorta for cardiac surgical guidance.
  • Embodiments of the present invention track motion the tip of a pigtail catheter, which is the loose circle at the distal end of the pigtail catheter.
  • Image (a) of FIG. 1 shows a pigtail catheter tip 102 detected in a frame of a fluoroscopic image sequence.
  • an annulus line can be identified in a 2D fluoroscopic image for visual guidance of the TAVI procedure.
  • Image (b) of FIG. 1 shows an annulus line 104 identified in a 2D fluoroscopic image when contrast in injected.
  • the pigtail catheter is continuously tracked in the 2D fluoroscopic images.
  • Images (c) and (d) respectively show tracked pigtail catheter tip locations 112 and 122 in two frames of a fluoroscopic image sequence, the location of the annulus line 104 identified when contrast was injected, motion-compensated annulus lines 114 and 124 determine based on the tracked pigtail catheter tip locations 112 and 122 , respectively. Accordingly, the catheter tracking can provide visual guidance for the aortic valve implantation, and also greatly reduce the amount of contrast agent used during interventions.
  • the annulus lines are used as examples, the present invention is not limited thereto. It is also possible that other geometric models, for example the aorta model, are used to visualize the motion compensation.
  • FIG. 2 illustrates examples of pigtail catheters in different fluoroscopic image sequences.
  • Image (a) shows a fluoroscopic image in which a pigtail catheter tip 202 that appears as a circle.
  • Image (b) shows a fluoroscopic images in which a pigtail catheter tip 204 that appears as an ellipse.
  • Image (c) shows a fluoroscopic image in which a pigtail catheter tip 206 appears as a straight line.
  • Image (d) shows a fluoroscopic image in which a pigtail catheter tip 208 is barely visible due to a low dose radiation x-ray and motion blurring.
  • a pigtail catheter also continuously moves with breathing motion and cardiac motion during image acquisition. As a result of these motions, a catheter could undergo rotation and twisting, and its shape and appearance can change within a fluoroscopic image sequence. Furthermore, when x-ray images are acquired during cardiac interventions, contrast agent is frequently injected in vessels in order to visualize the aorta, and a catheter may be occluded by the injected contrast agent. Moreover, in addition to the pigtail catheter, other devices, such as other catheters, stents, and probes, and anatomic structures, such as ribs and spines, may appear in the x-ray images.
  • FIG. 3 illustrates a sequence of fluoroscopic images in which a pigtail catheter is occluded by another device. As shown in FIG. 3 , in a sequence of fluoroscopic images with contrast injection, the pigtail catheter tip 302 is occluded by a stent 304 .
  • Embodiments of the present invention exploit rich and dynamic information available in fluoroscopic image sequences.
  • Embodiments of the present invention provide a framework, referred to herein as “discriminant tracking”, to obtain robust measurements modules to track a class of objects, e.g., pigtail catheters, in dynamic environments.
  • Embodiments of the present invention utilize an adaptive discriminant learning method to build online measurement models to discriminate objects from background.
  • a closed-form analytical solution is developed to efficiently update a discriminant function directly from online images without updating sample covariance matrices.
  • Embodiments of the present invention further introduce a Bayesian based fusion of multiple measurement models in a single measurement model. According to an advantageous embodiment three types of measurement models are combined, with each model exploiting a different type of information available in the image sequence:
  • the adaptive discriminant learning model utilizes a learning method to train a model to separate objects from backgrounds and can effectively handle dynamic environments that can change from one sequence to another.
  • the object class is denoted as the “positive” class represented ⁇ + by and the background is denoted as the “negative” class represented by ⁇ ⁇ .
  • An observed image is denoted as Z.
  • a data vector extracted from an image patch is denoted as x, and its associated class label is l x .
  • P(x) and P( ⁇ i ) are the prior probabilities of a data vector x and an object ⁇ i , respectively.
  • the transformation vector ⁇ k is the eigenvector corresponding to the k-th eigenvalue.
  • the rank of ⁇ h is 1, therefore only one effective eigenvector can be extracted from FDA.
  • the FDA transformation vector is denoted as A f .
  • a p ⁇ 1 , . . .
  • the sigmoid function may be more suitable to model the posterior probability of two class classifications:
  • Equation (5) provides a probabilistic model of linear discriminant analysis.
  • the probabilistic linear discriminant can be learned offline from collected training samples.
  • the background and sometimes the object itself, can change from one tracking scenario to another, and may not be seen in the training sets and therefore not be available for offline learning.
  • embodiments of the present invention provide an adaptive discriminant learning method for efficiently and effectively updating the discriminant analysis using online image data.
  • the probabilistic linear discriminant model is retrained online using a gradient descent method.
  • the discriminant model is iteratively updated as shown in Equations (7):
  • a (k) A (k ⁇ 1) ⁇ k ⁇ A C ( A (k ⁇ 1) ), (7)
  • ⁇ k is an updating step at the k-th iteration.
  • the updating is iterated until it converges.
  • the gradient of the object function ⁇ A C(A) is given by:
  • FIG. 4 illustrates an example of adaptive discriminant learning for detection of a pigtail catheter in a fluoroscopic image sequence.
  • the pigtail catheter appears as almost a line and the initial discriminant model has a large error of above 40%.
  • A is updated at each frame based on the tracked results (or the initialization at the first frame) as positive samples and image patches away from the tracked objects as negative samples.
  • Image (b) of FIG. 4 shows a curve representing the error of the discriminant model during adaptive learning at a frame.
  • the error of the initial discriminant model begins at above 40% and the greedy searching (i.e., the gradient descent iterations) converges within a few iterations to achieve a Bayesian error of less than 10%.
  • the greedy searching i.e., the gradient descent iterations
  • a history of tracked objects from previous frames is kept in the positive pools to make sure the learning is not distracted by tracking errors at individual frames.
  • the Bayesian error as expressed in Equation (6) is used as a criterion to decide if the learning has converged to an optimal solution. This allows the adaptive learned discriminant model to exclude some non-objects from positive samples that may be caused by occlusions or tracking errors.
  • FIG. 5 illustrates a method for tracking an object in a sequence of medical images according to an embodiment of the present invention.
  • the method of FIG. 5 may be used to track a pigtail catheter in a sequence of fluoroscopic images in a trans-catheter aortic valve implantation (TAVI).
  • TAVI trans-catheter aortic valve implantation
  • the present invention is no limited thereto, and the method of FIG. 5 may be similarly applied for tracking other medical devices as well.
  • the method of FIG. 5 utilizes a Bayesian framework for measurement fusion in object tracking.
  • the method of FIG. 5 integrates three types of measurements for tracking: an adaptively learned discriminant model, an offline learned object detection model, and an online appearance model. The models exploit different types of information, while their fusion makes the overall tracking more robust to dynamic changes of objects in a cluttered environment.
  • a first frame of a fluoroscopic image sequence is received.
  • the first frame of the fluoroscopic image sequence may be received from an x-ray imaging device as the fluoroscopic image sequence is captured by the x-ray imaging device.
  • the fluoroscopic image sequence is received by loading previously generated fluoroscopic images from a computer readable medium, or memory or storage of a computer system.
  • the fluoroscopic image sequence is a temporal series of frames, in which each frame is a 2D fluoroscopic image.
  • the fluoroscopic image sequence can be received by receiving fluoroscopic images from an x-ray imaging device in real time during an interventional procedure, such as a TAVI procedure.
  • the object is initialized in the first frame of the fluoroscopic image sequence.
  • the object e.g., pigtail catheter
  • the object may be initialized in a first frame by a user manually annotating the location of the pigtail catheter in the first frame.
  • the object may be initialized by automatically detecting the object in the first frame of the fluoroscopic image sequence.
  • a pigtail catheter can be automatically detected in a frame of a fluoroscopic image sequence using the method disclosed in U.S. Provisional Application No. 61/505,131, filed Jul. 7, 2011, the disclosure of which is incorporated herein by reference.
  • an adaptive discriminant model is updated based on the tracked object in the current frame.
  • online adaptive learning is applied to the discriminant model in order to update the discriminant model based on the tracked object in the current frame or the initialized object in the first frame.
  • an initial discriminant vector which is learned offline from annotated training data, is updated based on the initialized object.
  • FIG. 6 illustrates an algorithm for performing the adaptive discriminant learning according to an embodiment of the present invention. As illustrated in FIG.
  • a first stage 600 of the algorithm is performed offline to train an initial discriminant vector A (0) and a second stage 610 of the algorithm is performed for each frame of the fluoroscopic image sequence to use adaptive learning to update the discriminant vector A and corresponding probabilistic model based on the tracked object in each frame.
  • the PCA and FDA are trained from training samples.
  • the training samples are extracted from training data that is annotated with ground truth object locations.
  • PCA is applied to reduce the dimensionality of the data
  • the FDA is applied to the principal components in order to train a linear discriminant vector as shown in Equation (4) above.
  • a probabilistic model of the linear discriminant analysis is learned, as shown in Equation (5).
  • the parameters of the probabilistic model can be learned from the training samples using regression.
  • the second stage 610 of the algorithm of FIG. 6 is performed online during object tracking for each frame of the image sequence.
  • the second stage 610 of the algorithm of FIG. 6 is used to implement step 506 of the method of FIG. 5 .
  • positive samples corresponding to the tracked object in the current (or initialized object in the first frame) and negative samples away from the object are generated.
  • the image patch containing the tracked object in the current frame is added as a positive sample.
  • the tracked (or initialized) object can be shifted and rotated in order to generate a number of additional positive samples near the tracked object. A certain number of image patches more than a certain distance away from the tracked object in the current frame may be randomly selected as negative samples.
  • the discriminant vector A is updated using gradient descent, as shown in Equations (7) and (8).
  • the probabilistic model is updated based on the updated discriminant vector, as shown in Equation (5).
  • the Bayesian error C(A) is calculated for the updated discriminant vector using Equation (6). Steps 614 - 618 are repeated until the discriminant vector A converges or the Bayesian error C(A) is less than a threshold (e.g., 0.1). Although not shown in FIG. 6 , it is also possible that steps 614 - 618 are repeated until a maximum number of iterations is reached.
  • the updated discriminant vector and probabilistic model are stored and the tracked objects in the current frame are saved as positive training samples. Accordingly, each frame for which the algorithm is repeated adds additional positive training samples corresponding to the tracked object in that frame.
  • the next frame of the fluoroscopic image sequence is received.
  • the next frame may be received directly from an image acquisition device.
  • the next frame can be received and processed in real-time during a TAVI procedure. It is also possible that the next frame is received by loading the next frame from a fluoroscopic image sequence stored in a storage or memory of a computer system. At this point, the received next frame is considered to be the current frame.
  • the object is tracked in the current frame based on the adaptive discriminant model, a trained object detector, and an online appearance model using measurement fusion.
  • the tracking of the object infers unknown object states, e.g., the catheter motion denoted as m t , from an observed image sequence Z 1:t .
  • Embodiments of the present invention formalize the parameter inference in a sequential Bayesian framework.
  • Z 1:t-1 ) can be expressed as: P ( m t
  • Z 1:t-1 ) ⁇ P ( m t
  • Z 1:t-1 ) is the motion prior probability which is propagated from previous frames of the fluoroscopic image sequence.
  • m t-1 ) is a dynamic model.
  • the dynamic probability is modeled as a Gaussian model, i.e., P(m t
  • m t-1 ) G(m t ;m t-1 , ⁇ m ), where G is a Gaussian distribution with the mean of m t-1 and the covariance matrix of ⁇ m .
  • the Gaussian prior model is used as the dynamic model due to its simplicity and effectiveness of imposing smoothness constraints for 2D motions.
  • the Gaussian dynamic model gives a greater probability to objects having a state that is close to the object tracked in the previous frame.
  • the tracking result for an object in a frame is the motion parameter corresponding to the maximal posterior probability:
  • the measurement model plays a critical role, as it directly relates observed images to object states to be inferred. However, it is difficult for one measurement model to handle dynamic changes of objects and cluttered background.
  • m t ) is a fusion of multiple measurement models.
  • the adaptive discriminant model learned online is denoted as P (1) (Z t
  • m t ) P A (x m t ), in equation x m , as in equation (5), where x. is the data vector extracted from the observed images based on the object state m t .
  • the other two measurement models are from an offline trained object detector and an online appearance model.
  • an object detector can be trained offline based on a set of annotated training samples. Such a trained object detector is capable of modeling objects with large variations and also may be robust to background noise.
  • a probabilistic boosting tree (PBT) can be used to train the object detector.
  • PBT is a tree based generalized form of AdaBoost classifiers and can effectively model a complex distribution of a class of objects.
  • the object detector can be trained using Haar features.
  • the outputs of the trained PBT object detector can be interpreted into probabilistic measurements for input image patches, as expressed in equation (13):
  • PBT Probabilistic classifiers
  • the present invention is not limited thereto. It is also possible, that other probabilistic classifiers are used as the offline learning based measurement model, for example, SVM, k-NN, and Random forest.
  • the online appearance based measurement is used to model individual object appearance, and to keep tracking the appearance changes of an object.
  • the appearance based measurement model aims to model the online appearance of an individual object, instead of the whole class of objects.
  • the online appearance model is for the representation of the object, instead of the separation of an object from other portions of an image.
  • the online appearance model takes the form of: P (3) ( Z t
  • D(x m t ) is a robust distance function that calculates differences between current observations x m t and an appearance template x 0 , given by:
  • ⁇ s k ⁇ is a set of pixels in the image template.
  • Each pixel in the template is assigned a weight, c k , to de-emphasize the pixels far away from the template center.
  • is a robust function to remove outliers:
  • the template x 0 is updated by the tracking results to keep up with dynamic object changes.
  • the template updating may be sensitive to image noises and prone to tracking “drifting”. While it is possible to use a more sophisticated template update strategy, the present inventors have found that the simple template updating method, when fusing with other measurements, works well in a cluttered background. This further demonstrates the benefits of multiple measurement fusion.
  • m t ) are used in equation (12) to detect the state of the object in a frame of the image sequence.
  • the measurement models can be used in equation (12) to detect image patches where the object is located in a frame.
  • the weights of the three measurement models can be set experimentally by a user.
  • a kernel-based multi-resolution method may be used.
  • the multi-resolution tracking measurements are made at a set of sample object states m t,j s , instead of the whole state space.
  • m t,j s are uniformly sampled in a 2D motion space.
  • the Markov conditional independence can be assumed that the observations at sampling points m t,j s are independent of the un-sampled points m t , i.e., P(Z t
  • m t , m j s ) P(Z t
  • m t ) ⁇ j ⁇ ⁇ P ⁇ ( Z t
  • the kernel-base measurement estimation can obtain smooth measurements from a set of samples, so as to reduce computations of measurements. By incrementally decreasing the sampling steps and the ⁇ s , coarse to fine tracking can be achieved for a frame of an image sequence.
  • the object tracking results for the current frame are output.
  • the object tracking results may be output by displaying the objecting tracking results in the current frame of the fluoroscopic image sequence.
  • the object tracking results for the current frame can be displayed on a display of a computer system. It is also possible that the object tracking results are output by storing the output tracking results, for example in a storage or memory of a computer system.
  • step 514 it is determined whether the current frame is the final frame in the fluoroscopic image sequence. If the current frame is not the final frame, the method returns to step 506 . Accordingly, steps 506 - 512 are repeated to track the object and output the tracking results in each subsequent frame. This results in the online discriminant model being adaptively updated as the object is tracked in each frame, and the object being tracked using the adaptive discriminant model, the trained object detector measurement model, and the online appearance model in each subsequent frame. This also results the object tracking results being output for each frame in real time as the object is tracked in each frame. If the current frame is the final frame at step 514 , the method proceeds to step 516 . At step 516 , the method ends.
  • the method of FIG. 5 illustrates an advantageous embodiment in which the object tracking results for each frame are output in real time as the object is tracked in each frame
  • the present invention is not limited thereto. It is also possible that the tracked results for the entire image sequence can be output together after the object is tracked in all of the frames of the fluoroscopic image sequence.
  • FIG. 7 illustrates exemplary pigtail catheter tracking results using the method of FIG. 5 .
  • Sequence (a) of FIG. 7 shows tracked pigtail catheter tip locations 702 , 704 , 706 , 708 , and 710 are in frames 701 , 703 , 705 , 707 , and 709 , respectively.
  • Sequence (a) shows the discriminant tracking can handle occlusion and contrast injection.
  • Sequence (b) of FIG. 7 shows tracked pigtail catheter tip locations 712 , 714 , 716 , 718 , and 720 in frames 711 , 713 , 715 , 717 , and 719 , respectively.
  • sequence (b) shows that the robust tracking can recover from tracking error in single frames.
  • the set of images in (c) of FIG. 7 shows additional pigtail catheter tip tracking results 722 , 724 , 726 , 728 , 730 , 732 , 734 , and 736 in frames 721 , 723 , 725 , 727 , 729 , 731 , 733 , and 735 , respectively, of other fluoroscopic image sequences.
  • Computer 802 contains a processor 804 which controls the overall operation of the computer 802 by executing computer program instructions which define such operation.
  • the computer program instructions may be stored in a storage device 812 , or other computer readable medium (e.g., magnetic disk, CD ROM, etc.) and loaded into memory 810 when execution of the computer program instructions is desired.
  • An image acquisition device 820 such as an x-ray scanning device, can be connected to the computer 802 to input images to the computer 802 . It is possible to implement the image acquisition device 820 and the computer 802 as one device. It is also possible that the image acquisition device 820 and the computer 802 communicate wirelessly through a network.
  • the computer 802 also includes one or more network interfaces 806 for communicating with other devices via a network.
  • the computer 802 also includes other input/output devices 808 that enable user interaction with the computer 802 (e.g., display, keyboard, mouse, speakers, buttons, etc.).
  • FIG. 8 is a high level representation of some of the components of such a computer for illustrative purposes.

Abstract

A method and system for adaptive discriminant learning and measurement fusion for image based catheter tracking is disclosed. An adaptive discriminant model is trained online based on a tracked object, such as a pigtail catheter tip, in at least one previous frame of a fluoroscopic image sequence. The object is tracked in the current frame of the fluoroscopic image sequence based at least on the adaptive discriminant model trained online. The object may be tracked in the current frame based on a fusion of three types of measurement models including the adaptive discriminant model trained online, an object detection model trained offline, and an online appearance model.

Description

This application claims the benefit of U.S. Provisional Application No. 61/452,263, filed Mar. 14, 2011, and U.S. Provisional Application No. 61/505,131, filed Jul. 7, 2011, the disclosures of which are herein incorporated by reference.
BACKGROUND OF THE INVENTION
The present invention relates to medical imaging, and more particularly, to automatic tracking of catheter motion in a fluoroscopic image sequence.
Aortic valve disease affects a large number of people globally and is the most common type of valvular disease in developed countries. Implantation of a prosthetic aortic valve is often necessary to replace a severely damaged native valve. Although open-chest valve surgery is a well established procedure, minimally invasive transcatheter aortic valve implantation (TAVI) is an emerging technique, especially for high-risk patients, to minimize the surgical trauma. Interventional surgeries, such as TAVI, are typically performed under the guidance of real time fluoroscopic (x-ray) images. As the minimally invasive TAVI technique is emerging, physicians increasingly focus on minimizing the risks and making the surgery lass invasive in order to minimize the trauma, especially for high-risk patients. For example, it is desirable to reduce the times of exposure to and the amount of the potentially toxic contrast agent that is injected into a patient's blood. Most of such contrast in used to highlight the aorta and coronaries in fluoroscopic images in order to visually guide physicians. For example, when contrast is injected in TAVI, the aorta will be visible and an annulus line can be identified in a 2D fluoroscopic image. When there is no contrast injection, the aorta and annulus line will not be visible. Accordingly, it is desirable to track the motion of the aorta in non-contrast enhanced fluoroscopic images in order to reduce the exposure of patients to the contrast agent.
BRIEF SUMMARY OF THE INVENTION
The present invention provides a method and system for automatically tracking an object, such as a catheter, in a sequence of fluoroscopic images. Embodiments of the present invention utilize adaptive discriminant learning and measurement fusion for image-based catheter tracking. Embodiments of the present invention can be used to track a pigtail catheter in trans-catheter aortic valve implantation (TAVI). Since the pigtail catheter has the same motion as the aorta, it is possible to track the motion of the aorta by tracking the pigtail catheter.
In one embodiment of the present invention, an adaptive discriminant model is trained online based on a tracked object, such as a pigtail catheter tip, in at least one previous frame of a fluoroscopic image sequence. The object is tracked in the current frame of the fluoroscopic image sequence based at least on the adaptive discriminant model trained online. The object may be tracked in the current frame based on a fusion of three types of measurement models including the adaptive discriminant model trained online, an object detection model trained offline, and an online appearance model.
These and other advantages of the invention will be apparent to those of ordinary skill in the art by reference to the following detailed description and the accompanying drawings.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 illustrates an example of motion compensation for aortic valve implantation by tracking a pigtail catheter;
FIG. 2 illustrates examples of pigtail catheters in different fluoroscopic image sequences;
FIG. 3 illustrates a sequence of fluoroscopic images in which a pigtail catheter is occluded by another device;
FIG. 4 illustrates an example of adaptive discriminant learning for detection of a pigtail catheter in a fluoroscopic image sequence;
FIG. 5 illustrates a method for tracking an object in a sequence of medical images according to an embodiment of the present invention;
FIG. 6 illustrates an algorithm for performing the adaptive discriminant learning according to an embodiment of the present invention;
FIG. 7 illustrates exemplary pigtail catheter tracking results; and
FIG. 8 is a high level block diagram of a computer capable of implementing the present invention.
DETAILED DESCRIPTION
The present invention is directed to a method and system for automatically tracking an object, such as a catheter, in a fluoroscopic image sequence. A digital image is often composed of digital representations of one or more objects (or shapes). The digital representation of an object is often described herein in terms of identifying and manipulating the objects. Such manipulations are virtual manipulations accomplished in the memory or other circuitry/hardware of a computer system. Accordingly, it is to be understood that embodiments of the present invention may be performed within a computer system using data stored within the computer system.
Embodiments of the present invention provide a computational framework for robustly tracking an object by adaptive discriminant learning and measurement fusion. Such a tracking framework can be applied in Trans-catheter Aortic Valve Implantation (TAVI) in order to track a pigtail catheter. A medical device, such as a pigtail catheter can be tracked in a TAVI procedure to provide motion compensation for a 2D/3D overlay and reduce the exposure of a patient to contrast agent. FIG. 1 illustrates an example of motion compensation for aortic valve implantation by tracking a pigtail catheter. A pigtail catheter is a type of medical device that is inserted into the aorta for cardiac surgical guidance. Embodiments of the present invention track motion the tip of a pigtail catheter, which is the loose circle at the distal end of the pigtail catheter. Image (a) of FIG. 1 shows a pigtail catheter tip 102 detected in a frame of a fluoroscopic image sequence.
When a contrast agent is injected, the aorta and aortic valve will be visible for short period of time, an annulus line can be identified in a 2D fluoroscopic image for visual guidance of the TAVI procedure. Image (b) of FIG. 1 shows an annulus line 104 identified in a 2D fluoroscopic image when contrast in injected. When there is no contrast injection, aorta and aortic valve, and thus the annulus line, will not be visible. During the intervention, the pigtail catheter is continuously tracked in the 2D fluoroscopic images. Under the condition that the pigtail catheter moves with the same motion as the aorta, i.e., the pigtail catheter is not pulled during the intervention, the tracking of the pigtail catheter will provide motion compensation for continuous visualization of the annulus line. Images (c) and (d) respectively show tracked pigtail catheter tip locations 112 and 122 in two frames of a fluoroscopic image sequence, the location of the annulus line 104 identified when contrast was injected, motion-compensated annulus lines 114 and 124 determine based on the tracked pigtail catheter tip locations 112 and 122, respectively. Accordingly, the catheter tracking can provide visual guidance for the aortic valve implantation, and also greatly reduce the amount of contrast agent used during interventions. Although the annulus lines are used as examples, the present invention is not limited thereto. It is also possible that other geometric models, for example the aorta model, are used to visualize the motion compensation.
The dynamic clinical environment of the aortic valve implantation presents significant real-world problems for object tracking. For example, when x-ray (fluoroscopic) images are acquired at arbitrary angles, a catheter can appear as a circle, an ellipse, or even a straight line, depending on the projection angle. FIG. 2 illustrates examples of pigtail catheters in different fluoroscopic image sequences. Image (a) shows a fluoroscopic image in which a pigtail catheter tip 202 that appears as a circle. Image (b) shows a fluoroscopic images in which a pigtail catheter tip 204 that appears as an ellipse. Image (c) shows a fluoroscopic image in which a pigtail catheter tip 206 appears as a straight line. Image (d) shows a fluoroscopic image in which a pigtail catheter tip 208 is barely visible due to a low dose radiation x-ray and motion blurring.
A pigtail catheter also continuously moves with breathing motion and cardiac motion during image acquisition. As a result of these motions, a catheter could undergo rotation and twisting, and its shape and appearance can change within a fluoroscopic image sequence. Furthermore, when x-ray images are acquired during cardiac interventions, contrast agent is frequently injected in vessels in order to visualize the aorta, and a catheter may be occluded by the injected contrast agent. Moreover, in addition to the pigtail catheter, other devices, such as other catheters, stents, and probes, and anatomic structures, such as ribs and spines, may appear in the x-ray images. Some of the other devices and anatomical structures may be close to or even overlap with the pigtail catheter, which can cause automatic tracking of the pigtail catheter to fail. Because a low dose of radiation is preferred, x-ray images usually have a low signal-to-noise ratio, which can result in the catheter appearing blurry in an x-ray image, such as in image (d) of FIG. 2. FIG. 3 illustrates a sequence of fluoroscopic images in which a pigtail catheter is occluded by another device. As shown in FIG. 3, in a sequence of fluoroscopic images with contrast injection, the pigtail catheter tip 302 is occluded by a stent 304.
In order to address the above described difficulties in such dynamic environments, embodiments of the present invention exploit rich and dynamic information available in fluoroscopic image sequences. Embodiments of the present invention provide a framework, referred to herein as “discriminant tracking”, to obtain robust measurements modules to track a class of objects, e.g., pigtail catheters, in dynamic environments. Embodiments of the present invention utilize an adaptive discriminant learning method to build online measurement models to discriminate objects from background. A closed-form analytical solution is developed to efficiently update a discriminant function directly from online images without updating sample covariance matrices. Embodiments of the present invention further introduce a Bayesian based fusion of multiple measurement models in a single measurement model. According to an advantageous embodiment three types of measurement models are combined, with each model exploiting a different type of information available in the image sequence:
    • 1. An online adaptive discriminant model. The online adaptive discriminant model is adaptively learned for each sequence to separate an object from background;
    • 2. An offline learned object detection model. The object detection model is learned offline from a collected set of training data. The offline learned object detection model can handle large variations within a class of objects and can be generalized to a variety of environments; and
    • 3. An online appearance model. The online appearance model is built specifically for the object to be tracked. It exploits the consistency of the object between successive frames of an image sequence.
The adaptive discriminant learning model utilizes a learning method to train a model to separate objects from backgrounds and can effectively handle dynamic environments that can change from one sequence to another. For clarity of notations, the object class is denoted as the “positive” class represented Ω+ by and the background is denoted as the “negative” class represented by Ω. An observed image is denoted as Z. A data vector extracted from an image patch is denoted as x, and its associated class label is lx. lx=1 indicates that x belongs to the object class, while lx=−1 indicates that x belongs to the background.
In linear discriminant analysis, an original data vector is projected to a subspace of lower dimensionality where objects are more easily separated from the background. The projection is represented by a linear transformation, i.e., y=ΦTx, where Φ is the linear transformation matrix (or a vector). In an advantageous implementation, Fisher discriminant analysis (FDA) is used. FDA provides an optimal solution under the assumption that each class can be approximated by a Gaussian class distribution. FDA extracts discriminant features by maximizing the ratio of between-class and within-class variances, as shown in Equation (1):
J = Φ T Σ b Φ Φ T Σ w Φ , ( 1 )
where Σh and Σw are between-class and within-class scatter matrices, respectively, calculated as:
Σ w = i x Ω i P ( x ) ( x - x _ i ) ( x - x _ i ) T Σ b = i P ( Ω i ) ( x _ i - x _ ) ( x _ i - x _ ) T , ( 2 )
where x i is the mean vector of the samples from class Ωi and x is the mean vector of all samples. P(x) and P(Ωi) are the prior probabilities of a data vector x and an object Ωi, respectively. The FDA can be solved as a generalized eigenvector problem:
w −1ΣbkkΦk.  (3)
The transformation vector Φk is the eigenvector corresponding to the k-th eigenvalue. For the two class problem in object tracking, the rank of Σh is 1, therefore only one effective eigenvector can be extracted from FDA. The FDA transformation vector is denoted as Af.
In visual tracking, the data vector has high dimensionality. For example, an image data vector has more than 1000 elements for a 40 by 40 image patch, even after removing the corner pixels. The high dimensionality poses challenges to statistical learning as it requires a large amount of training samples. Principal component analysis (PCA) is applied before FDA to reduce data dimensionality. In PCA, each transformation vector Ψk corresponds to an eigenvector of a covariance matrix Σ+xεΩ + (x− x +)(x− x +)T, where x + is the mean of the object class. For the compact representation of the object class, only the first several significant principal components are selected, i.e., Ap=└Ψ1, . . . , Ψn┘ where n is much smaller than the original data dimension. After the PCA dimensionality reduction, the FDA is applied to principal components instead of the original data. The overall linear discriminant analysis is the combination of PCA and FDA, as shown in Equation (4):
y x =A f T A p T x=A T x,  (4)
where A=ApAf is the overall linear discriminant vector.
After learning the linear discriminant vector A, it is straightforward to model the projection of each class as a Gaussian distribution, from which the posterior probability can be estimated. However, according to an advantageous implementation, the sigmoid function may be more suitable to model the posterior probability of two class classifications:
P A ( x ) = P ( l x = 1 | x ) = 1 1 + ( ay x + b ) = 1 1 + ( aA T x + b ) . ( 5 )
The parameters a and b can be learned from training samples by regression. Equation (5) provides a probabilistic model of linear discriminant analysis.
The probabilistic linear discriminant can be learned offline from collected training samples. However, the background, and sometimes the object itself, can change from one tracking scenario to another, and may not be seen in the training sets and therefore not be available for offline learning. In dynamic environments, it is advantageous for the discriminant learning to adapt to different tracking scenarios. For this purpose, embodiments of the present invention provide an adaptive discriminant learning method for efficiently and effectively updating the discriminant analysis using online image data.
Most statistical learning algorithms aim at minimization of the Bayesian error, which is equivalent to minimizing the object function C(A) in Equation (6):
C ( A ) = - [ x i Ω i P ( x i ) P A ( x i ) - x j Ω - P ( x j ) P A ( x j ) ] = x - l x P ( x ) 1 + ( aA t x + b ) . ( 6 )
With new online data available during tracking, re-training the probabilistic linear discriminant model may require a large number of training samples and involve re-computation of co-variance metrics and eigenbases. In an advantageous embodiment of the present invention, the probabilistic linear discriminant model is retrained online using a gradient descent method. To minimize the Bayesian error, the discriminant model is iteratively updated as shown in Equations (7):
A (k) =A (k−1)−δkA C(A (k−1)),  (7)
Where δk is an updating step at the k-th iteration. The updating is iterated until it converges. The gradient of the object function ∇AC(A) is given by:
A C ( A ) = x 1 A ( - i P ( x i ) 1 + ( aA t t x t + b ) ) = x t l i P ( x i ) a ( aA t x t + b ) ( 1 + ( aA t x t + b ) ) 2 x i . ( 8 )
After the linear discriminant vector A is updated at each iteration, the parameters a and b are also updated via regression.
Since the adaptive learning is a greedy method, it needs a good starting point to converge to a good solution. The starting point can be the discriminant learned offline from collected training samples, denoted as A(0). Even if the initial discriminant does not fit the current environment, the adaptive learning can quickly converge to a good solution. FIG. 4 illustrates an example of adaptive discriminant learning for detection of a pigtail catheter in a fluoroscopic image sequence. In the example of FIG. 4, the pigtail catheter appears as almost a line and the initial discriminant model has a large error of above 40%. During tracking, A is updated at each frame based on the tracked results (or the initialization at the first frame) as positive samples and image patches away from the tracked objects as negative samples. Image (a) of FIG. 4 shows positive samples 402 and negative samples 404 extracted from a frame of a fluoroscopic image sequence. In a possible implementation, the tracked (or initialized) object can be shifted and rotated to introduce a variation in the positive samples 402. Image (b) of FIG. 4 shows a curve representing the error of the discriminant model during adaptive learning at a frame. As shown in image (b) of FIG. 4, the error of the initial discriminant model begins at above 40% and the greedy searching (i.e., the gradient descent iterations) converges within a few iterations to achieve a Bayesian error of less than 10%. To improve learning robustness, a history of tracked objects from previous frames is kept in the positive pools to make sure the learning is not distracted by tracking errors at individual frames. Furthermore, the Bayesian error as expressed in Equation (6) is used as a criterion to decide if the learning has converged to an optimal solution. This allows the adaptive learned discriminant model to exclude some non-objects from positive samples that may be caused by occlusions or tracking errors.
FIG. 5 illustrates a method for tracking an object in a sequence of medical images according to an embodiment of the present invention. In an advantageous embodiment, the method of FIG. 5 may be used to track a pigtail catheter in a sequence of fluoroscopic images in a trans-catheter aortic valve implantation (TAVI). However, the present invention is no limited thereto, and the method of FIG. 5 may be similarly applied for tracking other medical devices as well. The method of FIG. 5 utilizes a Bayesian framework for measurement fusion in object tracking. The method of FIG. 5 integrates three types of measurements for tracking: an adaptively learned discriminant model, an offline learned object detection model, and an online appearance model. The models exploit different types of information, while their fusion makes the overall tracking more robust to dynamic changes of objects in a cluttered environment.
As illustrated in FIG. 5, at step 502, a first frame of a fluoroscopic image sequence is received. For example, the first frame of the fluoroscopic image sequence may be received from an x-ray imaging device as the fluoroscopic image sequence is captured by the x-ray imaging device. It is also possible that the fluoroscopic image sequence is received by loading previously generated fluoroscopic images from a computer readable medium, or memory or storage of a computer system. The fluoroscopic image sequence is a temporal series of frames, in which each frame is a 2D fluoroscopic image. According to an advantageous embodiment, the fluoroscopic image sequence can be received by receiving fluoroscopic images from an x-ray imaging device in real time during an interventional procedure, such as a TAVI procedure.
At step 504, the object is initialized in the first frame of the fluoroscopic image sequence. In one embodiment, the object, e.g., pigtail catheter, may be initialized in a first frame by a user manually annotating the location of the pigtail catheter in the first frame. In an alternative embodiment, the object may be initialized by automatically detecting the object in the first frame of the fluoroscopic image sequence. For example, a pigtail catheter can be automatically detected in a frame of a fluoroscopic image sequence using the method disclosed in U.S. Provisional Application No. 61/505,131, filed Jul. 7, 2011, the disclosure of which is incorporated herein by reference.
At step 506, an adaptive discriminant model is updated based on the tracked object in the current frame. In particular, online adaptive learning is applied to the discriminant model in order to update the discriminant model based on the tracked object in the current frame or the initialized object in the first frame. In the first frame, an initial discriminant vector, which is learned offline from annotated training data, is updated based on the initialized object. FIG. 6 illustrates an algorithm for performing the adaptive discriminant learning according to an embodiment of the present invention. As illustrated in FIG. 6, a first stage 600 of the algorithm is performed offline to train an initial discriminant vector A(0) and a second stage 610 of the algorithm is performed for each frame of the fluoroscopic image sequence to use adaptive learning to update the discriminant vector A and corresponding probabilistic model based on the tracked object in each frame. In the first stage 600, which is performed offline, at 602, the PCA and FDA are trained from training samples. The training samples are extracted from training data that is annotated with ground truth object locations. In particular, PCA is applied to reduce the dimensionality of the data, and after the PCA dimensionality reduction, the FDA is applied to the principal components in order to train a linear discriminant vector as shown in Equation (4) above. At 604, a probabilistic model of the linear discriminant analysis is learned, as shown in Equation (5). The parameters of the probabilistic model can be learned from the training samples using regression.
The second stage 610 of the algorithm of FIG. 6 is performed online during object tracking for each frame of the image sequence. In an advantageous implementation, the second stage 610 of the algorithm of FIG. 6 is used to implement step 506 of the method of FIG. 5. At 612, positive samples corresponding to the tracked object in the current (or initialized object in the first frame) and negative samples away from the object are generated. In particular, the image patch containing the tracked object in the current frame is added as a positive sample. According to an advantageous implementation, the tracked (or initialized) object can be shifted and rotated in order to generate a number of additional positive samples near the tracked object. A certain number of image patches more than a certain distance away from the tracked object in the current frame may be randomly selected as negative samples. At 614, the discriminant vector A is updated using gradient descent, as shown in Equations (7) and (8). At 616, the probabilistic model is updated based on the updated discriminant vector, as shown in Equation (5). At 618, The Bayesian error C(A) is calculated for the updated discriminant vector using Equation (6). Steps 614-618 are repeated until the discriminant vector A converges or the Bayesian error C(A) is less than a threshold (e.g., 0.1). Although not shown in FIG. 6, it is also possible that steps 614-618 are repeated until a maximum number of iterations is reached. At 620, when the learning converges, the updated discriminant vector and probabilistic model are stored and the tracked objects in the current frame are saved as positive training samples. Accordingly, each frame for which the algorithm is repeated adds additional positive training samples corresponding to the tracked object in that frame.
Returning to FIG. 5, at step 508, the next frame of the fluoroscopic image sequence is received. The next frame may be received directly from an image acquisition device. For example, the next frame can be received and processed in real-time during a TAVI procedure. It is also possible that the next frame is received by loading the next frame from a fluoroscopic image sequence stored in a storage or memory of a computer system. At this point, the received next frame is considered to be the current frame.
At step 510, the object is tracked in the current frame based on the adaptive discriminant model, a trained object detector, and an online appearance model using measurement fusion. The tracking of the object infers unknown object states, e.g., the catheter motion denoted as mt, from an observed image sequence Z1:t. Embodiments of the present invention formalize the parameter inference in a sequential Bayesian framework. Assuming a commonly used Markov property for tracking, the posterior probability P(mt|Z1:t-1) can be expressed as:
P(m t |Z 1:t-1)=∫P(m t |m t-1)P(m t-1 |Z 1:t-1)dm t-1 P(m t |Z 1:t)∝P(m t |Z 1:t-1)P(Z t |m t).  (9)
In Equation (9), P(mt|Z1:t-1) is the motion prior probability which is propagated from previous frames of the fluoroscopic image sequence. P(mt|mt-1) is a dynamic model. In an advantageous implementation, the dynamic probability is modeled as a Gaussian model, i.e., P(mt|mt-1)=G(mt;mt-1m), where G is a Gaussian distribution with the mean of mt-1 and the covariance matrix of Σm. The Gaussian prior model is used as the dynamic model due to its simplicity and effectiveness of imposing smoothness constraints for 2D motions. The Gaussian dynamic model gives a greater probability to objects having a state that is close to the object tracked in the previous frame. Another component, the likelihood measurement model P(Zt|mt), measures the likelihood of motion parameters. The tracking result for an object in a frame is the motion parameter corresponding to the maximal posterior probability:
m ^ t = arg m t max P ( m t | Z 1 : t ) . ( 10 )
In visual tracking, the measurement model plays a critical role, as it directly relates observed images to object states to be inferred. However, it is difficult for one measurement model to handle dynamic changes of objects and cluttered background. According to an advantageous aspect of the present invention, the measurement model P(Zt|mt) is a fusion of multiple measurement models.
Assuming that there are K measurement models for an object, i.e., P(k)(Zt|mt)=P(Zt|mt,k), k=1, . . . , K, a Bayesian measurement fusion based on the probability marginalization can be expressed as:
P ( Z t | m t ) = k P ( Z t | m t , k ) P ( k | m t ) . ( 11 )
For simplicity, it can be assumed that the measurement model prior P(k|mt) does not change with object states during tracking, i.e., P(k|mt)=ωk, and Σkωk=1. ωk denotes a weight of the k-th measurement model. Therefore, an overall measurement model can be expressed as P(Zt|mt)=ΣkωkP(k)(Zt|mt), and equation (9) can be re-written as:
P ( m t | Z 1 : t ) P ( m t | Z 1 : t - 1 ) k ω k P ( k ) ( Z t | m t ) . ( 12 )
In an advantageous embodiment of the present invention, three types of measurement models are fused. The first is the adaptive discriminant model learned online (at step 506), as described above. The adaptive discriminant model is denoted as P(1)(Zt|mt)=PA(xm t ), in equation xm, as in equation (5), where x. is the data vector extracted from the observed images based on the object state mt. The other two measurement models are from an offline trained object detector and an online appearance model.
In order to provide the offline learning based measurement model, an object detector can be trained offline based on a set of annotated training samples. Such a trained object detector is capable of modeling objects with large variations and also may be robust to background noise. In an advantageous implementation, a probabilistic boosting tree (PBT) can be used to train the object detector. A PBT is a tree based generalized form of AdaBoost classifiers and can effectively model a complex distribution of a class of objects. The object detector can be trained using Haar features. The output of a trained PBT classifier, denoted as f(x), is a combination of outputs from a collection of learned weak classifiers Hk(x) with associated weights αk, i.e., f(x)=ΣkαkHk(x). The outputs of the trained PBT object detector can be interpreted into probabilistic measurements for input image patches, as expressed in equation (13):
P ( 2 ) ( Z t | m t ) = f ( x m t ) - f ( x m t ) + f ( x m t ) . ( 13 )
Although PBT is used, the present invention is not limited thereto. It is also possible, that other probabilistic classifiers are used as the offline learning based measurement model, for example, SVM, k-NN, and Random forest.
The online appearance based measurement is used to model individual object appearance, and to keep tracking the appearance changes of an object. Different from the offline learned measurement model, the appearance based measurement model aims to model the online appearance of an individual object, instead of the whole class of objects. Also different from the online adaptive discriminant model, the online appearance model is for the representation of the object, instead of the separation of an object from other portions of an image. In an advantageous implementation, the online appearance model takes the form of:
P (3)(Z t |m t)∝G(D(x m t );σα),  (14)
where G is a one dimensional Gaussian kernel with a zero mean and a bandwidth of σα. D(xm t ) is a robust distance function that calculates differences between current observations xm t and an appearance template x0, given by:
D ( m t ) = k c k ρ ( x m t ( s k ) - x 0 ( s k ) ; σ a ) . ( 15 )
In equation (15), {sk} is a set of pixels in the image template. Each pixel in the template is assigned a weight, ck, to de-emphasize the pixels far away from the template center. The pixel weight is set as ck=G(−|sk|;σt), where |sk| is the distance for sk to the template center, and σt is set as the image template size. ρ is a robust function to remove outliers:
ρ ( y ; σ a ) = { y , if y <= 3 σ a 3 σ a , if y > 3 σ a . ( 16 )
During tracking, the template x0 is updated by the tracking results to keep up with dynamic object changes. The template updating may be sensitive to image noises and prone to tracking “drifting”. While it is possible to use a more sophisticated template update strategy, the present inventors have found that the simple template updating method, when fusing with other measurements, works well in a cluttered background. This further demonstrates the benefits of multiple measurement fusion.
The three measurement models P(1)(Zt|mt), P(2)(Zt|mt), and P(3)(Zt|mt) are used in equation (12) to detect the state of the object in a frame of the image sequence. For example, the measurement models can be used in equation (12) to detect image patches where the object is located in a frame. The weights of the three measurement models can be set experimentally by a user. In a possible implementation, the measurement models P(1)(Zt|mt), P(2)(Zt|mt), and P(3)(Zt|mt) can be assigned weights of ω1=0.25, ω2=0.50, and ω3=0.25, but the present invention is not limited thereto.
Exhaustively searching the object state space to obtain the maximal posterior probability estimation is computationally expensive. For computational efficiency, a kernel-based multi-resolution method may be used. In the multi-resolution tracking, measurements are made at a set of sample object states mt,j s, instead of the whole state space. In this method, mt,j s are uniformly sampled in a 2D motion space. The Markov conditional independence can be assumed that the observations at sampling points mt,j s are independent of the un-sampled points mt, i.e., P(Zt|mt, mj s)=P(Zt|mj s). Therefore, the kernel-based measurement estimation can be represented as:
P ( Z t | m t ) = j P ( Z t | m t , j a ) P ( m t , j a | m t ) ,
where P(mj,t s|mt)=G(mt,j ss) is a Gaussian kernel with a bandwidth σs. The kernel-base measurement estimation can obtain smooth measurements from a set of samples, so as to reduce computations of measurements. By incrementally decreasing the sampling steps and the σs, coarse to fine tracking can be achieved for a frame of an image sequence.
Returning to FIG. 5, at step 512, the object tracking results for the current frame are output. For example the object tracking results may be output by displaying the objecting tracking results in the current frame of the fluoroscopic image sequence. For example, the object tracking results for the current frame can be displayed on a display of a computer system. It is also possible that the object tracking results are output by storing the output tracking results, for example in a storage or memory of a computer system.
At step 514, it is determined whether the current frame is the final frame in the fluoroscopic image sequence. If the current frame is not the final frame, the method returns to step 506. Accordingly, steps 506-512 are repeated to track the object and output the tracking results in each subsequent frame. This results in the online discriminant model being adaptively updated as the object is tracked in each frame, and the object being tracked using the adaptive discriminant model, the trained object detector measurement model, and the online appearance model in each subsequent frame. This also results the object tracking results being output for each frame in real time as the object is tracked in each frame. If the current frame is the final frame at step 514, the method proceeds to step 516. At step 516, the method ends.
Although the method of FIG. 5 illustrates an advantageous embodiment in which the object tracking results for each frame are output in real time as the object is tracked in each frame, the present invention is not limited thereto. It is also possible that the tracked results for the entire image sequence can be output together after the object is tracked in all of the frames of the fluoroscopic image sequence.
FIG. 7 illustrates exemplary pigtail catheter tracking results using the method of FIG. 5. Sequence (a) of FIG. 7 shows tracked pigtail catheter tip locations 702, 704, 706, 708, and 710 are in frames 701, 703, 705, 707, and 709, respectively. Sequence (a) shows the discriminant tracking can handle occlusion and contrast injection. Sequence (b) of FIG. 7 shows tracked pigtail catheter tip locations 712, 714, 716, 718, and 720 in frames 711, 713, 715, 717, and 719, respectively. As shown in frame 715, there is an error between the tracked pigtail catheter tip location 716 and the ground truth pigtail catheter tip location 750. Accordingly, sequence (b) shows that the robust tracking can recover from tracking error in single frames. The set of images in (c) of FIG. 7 shows additional pigtail catheter tip tracking results 722, 724, 726, 728, 730, 732, 734, and 736 in frames 721, 723, 725, 727, 729, 731, 733, and 735, respectively, of other fluoroscopic image sequences.
The above-described methods for tracking an object, such as a pigtail catheter, in a fluoroscopic image sequence may be implemented on a computer using well-known computer processors, memory units, storage devices, computer software, and other components. A high level block diagram of such a computer is illustrated in FIG. 8. Computer 802 contains a processor 804 which controls the overall operation of the computer 802 by executing computer program instructions which define such operation. The computer program instructions may be stored in a storage device 812, or other computer readable medium (e.g., magnetic disk, CD ROM, etc.) and loaded into memory 810 when execution of the computer program instructions is desired. Thus, the steps of the methods of FIGS. 5 and 6 may be defined by the computer program instructions stored in the memory 810 and/or storage 812 and controlled by the processor 804 executing the computer program instructions. An image acquisition device 820, such as an x-ray scanning device, can be connected to the computer 802 to input images to the computer 802. It is possible to implement the image acquisition device 820 and the computer 802 as one device. It is also possible that the image acquisition device 820 and the computer 802 communicate wirelessly through a network. The computer 802 also includes one or more network interfaces 806 for communicating with other devices via a network. The computer 802 also includes other input/output devices 808 that enable user interaction with the computer 802 (e.g., display, keyboard, mouse, speakers, buttons, etc.). One skilled in the art will recognize that an implementation of an actual computer could contain other components as well, and that FIG. 8 is a high level representation of some of the components of such a computer for illustrative purposes.
The foregoing Detailed Description is to be understood as being in every respect illustrative and exemplary, but not restrictive, and the scope of the invention disclosed herein is not to be determined from the Detailed Description, but rather from the claims as interpreted according to the full breadth permitted by the patent laws. It is to be understood that the embodiments shown and described herein are only illustrative of the principles of the present invention and that various modifications may be implemented by those skilled in the art without departing from the scope and spirit of the invention. Those skilled in the art could implement various other feature combinations without departing from the scope and spirit of the invention.

Claims (26)

The invention claimed is:
1. A method for tracking an object in a sequence of fluoroscopic images, comprising:
training an adaptive discriminant model online based on a tracked object in at least one previous frame of the fluoroscopic image sequence; and
tracking the object in a current frame of the fluoroscopic image sequence based at least on the adaptive discriminant model trained online, wherein tracking the object in a current frame of the fluoroscopic image sequence based at least on the adaptive discriminant model trained online comprises:
tracking the object in the current frame using an overall measurement model that is a fusion of the adaptive discriminant model trained online, an object detection model trained offline, and an online appearance model.
2. The method of claim 1, wherein the step of training an adaptive discriminant model online based on a tracked object in at least one previous frame of the fluoroscopic image sequence comprises:
extracting positive samples from the tracked object in the at least one previous frame and negative samples away from the tracked object in the at least one previous frame;
updating a linear discriminant vector based on the extracted positive samples and negative samples using a gradient descent method to reduce a Bayesian error; and
updating a probabilistic model based on the updated linear discriminant vector.
3. The method of claim 2, wherein the step of training an adaptive discriminant model online based on a tracked object in at least one previous frame of the fluoroscopic image sequence further comprises:
repeating the steps of updating the linear discriminant vector and updating the probabilistic model until the linear discriminant vector converges.
4. The method of claim 2, wherein an initial linear discriminant vector is trained offline based on annotated training data.
5. The method of claim 4, wherein the initial linear discriminant vector is trained offline using principal component analysis (PCA) to reduce the dimensionality of the training data and Fisher discriminant analysis (FDA) applied to principal components result from the PCA to learn the initial discriminant vector.
6. The method of claim 1, wherein the object detection model is trained offline based on annotated training data using a probabilistic boosting tree (PBT).
7. The method of claim 1, wherein the online appearance model calculates a probabilities for image patches in the current frame based on differences between the image patches and an appearance template trained online from the tracked object in the at least one previous frame.
8. The method of claim 1, wherein the step of tracking the object in the current frame using an overall measurement model that is a fusion of the adaptive discriminant model trained online, an object detection model trained offline, and an online appearance model comprises:
tracking the object in the current frame based on a motion prior probability which is propagated from the at least one previous frame and the overall measurement model.
9. The method of claim 1, wherein the step of tracking the object in a current frame of the fluoroscopic image sequence based at least on the adaptive discriminant model trained online comprises:
detecting motion parameters mt in the object in the current frame t to maximize a posterior probability P(mt|Z1:t):
P ( m t | Z 1 : t ) P ( m t | Z 1 : t - 1 ) k = 1 3 ω k P ( k ) ( Z t | m t ) ,
where Zi is observed image data of frame i, P(mt|Z1:t-1) is a motion prior probability propagated from the at least one previous frame, P(t)(Zt|mt) is the adaptive discriminant model trained online, P(2)(Zt|mt) is a probabilistic measurement from an object detector trained offline, and P(3)(Zt|mt) is an online appearance model.
10. The method of claim 1, wherein the object is a pigtail catheter tip.
11. An apparatus for tracking an object in a sequence of fluoroscopic images, comprising:
means for training an adaptive discriminant model online based on a tracked object in at least one previous frame of the fluoroscopic image sequence; and
means for tracking the object in a current frame of the fluoroscopic image sequence based at least on the adaptive discriminant model trained online, comprising:
means for tracking the object in the current frame using an overall measurement model that is a fusion of the adaptive discriminant model trained online, an object detection model trained offline, and an online appearance model.
12. The apparatus of claim 11, wherein the means for training an adaptive discriminant model online based on a tracked object in at least one previous frame of the fluoroscopic image sequence comprises:
means for extracting positive samples from the tracked object in the at least one previous frame and negative samples away from the tracked object in the at least one previous frame;
means for updating a linear discriminant vector based on the extracted positive samples and negative samples using a gradient descent method to reduce a Bayesian error; and
means for updating a probabilistic model based on the updated linear discriminant vector.
13. The apparatus of claim 12, further comprising:
means for training an initial linear discriminant vector offline based on annotated training data.
14. The apparatus of claim 11, wherein the object detection model is trained offline based on annotated training data using a probabilistic boosting tree (PBT).
15. The apparatus of claim 11, wherein the online appearance model calculates a probabilities for image patches in the current frame based on differences between the image patches and an appearance template trained online from the tracked object in the at least one previous frame.
16. The apparatus of claim 11, wherein the means for tracking the object in the current frame using an overall measurement model that is a fusion of the adaptive discriminant model trained online, an object detection model trained offline, and an online appearance model comprises:
means for tracking the object in the current frame based on a motion prior probability which is propagated from the at least one previous frame and the overall measurement model.
17. The apparatus of claim 11, wherein the object is a pigtail catheter tip.
18. A non-transitory computer readable medium encoded with computer executable instructions for tracking an object in a sequence of fluoroscopic images, the computer executable instructions defining a method comprising:
training an adaptive discriminant model online based on a tracked object in at least one previous frame of the fluoroscopic image sequence; and
tracking the object in a current frame of the fluoroscopic image sequence based at least on the adaptive discriminant model trained online, wherein tracking the object in a current frame of the fluoroscopic image sequence based at least on the adaptive discriminant model trained online comprises:
tracking the object in the current frame using an overall measurement model that is a fusion of the adaptive discriminant model trained online, an object detection model trained offline, and an online appearance model.
19. The non-transitory computer readable medium of claim 18, wherein the step of training an adaptive discriminant model online based on a tracked object in at least one previous frame of the fluoroscopic image sequence comprises:
extracting positive samples from the tracked object in the at least one previous frame and negative samples away from the tracked object in the at least one previous frame;
updating a linear discriminant vector based on the extracted positive samples and negative samples using a gradient descent method to reduce a Bayesian error; and
updating a probabilistic model based on the updated linear discriminant vector.
20. The non-transitory computer readable medium of claim 19, wherein the step of training an adaptive discriminant model online based on a tracked object in at least one previous frame of the fluoroscopic image sequence further comprises:
repeating the steps of updating the linear discriminant vector and updating the probabilistic model until the linear discriminant vector converges.
21. The non-transitory computer readable medium of claim 19, wherein an initial linear discriminant vector is trained offline based on annotated training data.
22. The non-transitory computer readable medium of claim 18, wherein the object detection model is trained offline based on annotated training data using a probabilistic boosting tree (PBT).
23. The non-transitory computer readable medium of claim 18, wherein the online appearance model calculates a probabilities for image patches in the current frame based on differences between the image patches and an appearance template trained online from the tracked object in the at least one previous frame.
24. The non-transitory computer readable medium of claim 18, wherein the step of tracking the object in the current frame using an overall measurement model that is a fusion of the adaptive discriminant model trained online, an object detection model trained offline, and an online appearance model comprises:
tracking the object in the current frame based on a motion prior probability which is propagated from the at least one previous frame and the overall measurement model.
25. The non-transitory computer readable medium of claim 18, wherein the step of tracking the object in a current frame of the fluoroscopic image sequence based at least on the adaptive discriminant model trained online comprises:
detecting motion parameters mt in the object in the current frame t to maximize a posterior probability P(mt|Z1:t):
P ( m t | Z 1 : t ) P ( m t | Z 1 : t - 1 ) k = 1 3 ω k P ( k ) ( Z t | m t ) ,
where Zi is observed image data of frame i, P(mt|Z1:t-1) is a motion prior probability propagated from the at least one previous frame, P(1)(Zt|mt) is the adaptive discriminant model trained online, P(2)(Zt|mt) is a probabilistic measurement from an object detector trained offline, and P(3)(Zt|mt) is an online appearance model.
26. The non-transitory computer readable medium of claim 18, wherein the object is a pigtail catheter tip.
US13/412,882 2011-03-14 2012-03-06 Method and system for catheter tracking in fluoroscopic images using adaptive discriminant learning and measurement fusion Active 2033-09-01 US9014423B2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US13/412,882 US9014423B2 (en) 2011-03-14 2012-03-06 Method and system for catheter tracking in fluoroscopic images using adaptive discriminant learning and measurement fusion
CN201210120135.2A CN102697482B (en) 2011-03-14 2012-03-14 Self adaptation is used to differentiate the method and system that the conduit of study and measurement fusion is followed the tracks of

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US201161452263P 2011-03-14 2011-03-14
US201161505131P 2011-07-07 2011-07-07
US13/412,882 US9014423B2 (en) 2011-03-14 2012-03-06 Method and system for catheter tracking in fluoroscopic images using adaptive discriminant learning and measurement fusion

Publications (2)

Publication Number Publication Date
US20120238866A1 US20120238866A1 (en) 2012-09-20
US9014423B2 true US9014423B2 (en) 2015-04-21

Family

ID=46829008

Family Applications (1)

Application Number Title Priority Date Filing Date
US13/412,882 Active 2033-09-01 US9014423B2 (en) 2011-03-14 2012-03-06 Method and system for catheter tracking in fluoroscopic images using adaptive discriminant learning and measurement fusion

Country Status (1)

Country Link
US (1) US9014423B2 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10143438B2 (en) 2015-08-06 2018-12-04 Xiang Zhang System for 3D object modeling and tracking in X-ray imaging
US11823376B2 (en) 2018-05-16 2023-11-21 Benevis Informatics, Llc Systems and methods for review of computer-aided detection of pathology in images

Families Citing this family (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9424648B2 (en) 2011-07-07 2016-08-23 Siemens Aktiengesellschaft Method and system for device detection in 2D medical images
US9259200B2 (en) * 2012-10-18 2016-02-16 Siemens Aktiengesellschaft Method and system for obtaining a sequence of x-ray images using a reduced dose of ionizing radiation
US20160140727A1 (en) * 2013-06-17 2016-05-19 Aselsan Elektronik Sanayi Ve Ticaret Anonim Sirket I A method for object tracking
US10362943B2 (en) 2013-09-20 2019-07-30 Siemens Healthcare Gmbh Dynamic overlay of anatomy from angiography to fluoroscopy
CN104318539A (en) * 2014-10-16 2015-01-28 哈尔滨工程大学 Sonar image fusion method based on extended Piella framework
CN107249464B (en) * 2015-02-27 2021-03-12 西门子保健有限责任公司 Robust calcification tracking in fluorescence imaging
WO2016181318A1 (en) 2015-05-12 2016-11-17 Navix International Limited Lesion assessment by dielectric property analysis
US10278616B2 (en) 2015-05-12 2019-05-07 Navix International Limited Systems and methods for tracking an intrabody catheter
US11039888B2 (en) 2015-05-12 2021-06-22 Navix International Limited Calculation of an ablation plan
JP2018520718A (en) 2015-05-12 2018-08-02 ナヴィックス インターナショナル リミテッドNavix International Limited Contact quality evaluation by dielectric property analysis
CN105590328B (en) * 2015-12-07 2018-04-03 天津大学 Frame adaptive target tracking algorism based on rarefaction representation selectivity display model
EP3484362A1 (en) 2016-07-14 2019-05-22 Navix International Limited Characteristic track catheter navigation
WO2018092070A1 (en) 2016-11-16 2018-05-24 Navix International Limited Esophagus position detection by electrical mapping
WO2018092062A1 (en) 2016-11-16 2018-05-24 Navix International Limited Real-time display of tissue deformation by interactions with an intra-body probe
WO2018092071A1 (en) 2016-11-16 2018-05-24 Navix International Limited Estimators for ablation effectiveness
WO2018092063A1 (en) 2016-11-16 2018-05-24 Navix International Limited Real-time display of treatment-related tissue changes using virtual material
US11010983B2 (en) 2016-11-16 2021-05-18 Navix International Limited Tissue model dynamic visual rendering
JP6999895B2 (en) * 2017-07-06 2022-01-19 株式会社島津製作所 Radiation imaging device and radiological image detection method
JP7113447B2 (en) * 2018-03-12 2022-08-05 東芝エネルギーシステムズ株式会社 Medical Image Processing Apparatus, Treatment System, and Medical Image Processing Program
CN110728325A (en) * 2019-10-12 2020-01-24 太原科技大学 Variable voltage DR image fusion method based on support degree transformation

Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020085681A1 (en) 2000-12-28 2002-07-04 Jensen Vernon Thomas Method and apparatus for obtaining and displaying computed tomography images using a fluoroscopy imaging system
WO2003041057A2 (en) 2001-11-08 2003-05-15 The Johns Hopkins University System and method for robot targeting under flouroscopy based on image servoing
US6961608B2 (en) 2000-06-05 2005-11-01 Kabushiki Kaisha Toshiba Interventional MR imaging with detection and display of device position
US20050245814A1 (en) * 2004-04-28 2005-11-03 Sunnybrook & Women's College Health Sciences Centre Catheter tracking with phase information
US20080033452A1 (en) * 2004-03-03 2008-02-07 Deutsches Krebsforschungszentrum Incremental Real-Time Recording Of Tracked Instruments In Tubular Organ Structures Inside The Human Body
CN101221620A (en) 2007-12-20 2008-07-16 北京中星微电子有限公司 Human face tracing method
CN101371784A (en) 2007-08-21 2009-02-25 美国西门子医疗解决公司 Method and system for catheter detection and tracking in a fluoroscopic image sequence
CN101661559A (en) 2009-09-16 2010-03-03 中国科学院计算技术研究所 Digital image training and detecting methods
US20100157041A1 (en) * 2007-03-08 2010-06-24 Sync-Rx, Ltd. Automatic stabilization of an image stream of a moving organ
US20100239148A1 (en) 2009-03-18 2010-09-23 Siemens Corporation Method and System for Automatic Aorta Segmentation
US20100249579A1 (en) 2007-12-17 2010-09-30 Starks Daniel R Systems and methods for modeling both unobstructed and obstructed portions of a catheter
CN101908153A (en) 2010-08-21 2010-12-08 上海交通大学 Method for estimating head postures in low-resolution image treatment
US20110038517A1 (en) 2009-08-17 2011-02-17 Mistretta Charles A System and method for four dimensional angiography and fluoroscopy
US20110164035A1 (en) 2010-01-07 2011-07-07 Siemens Corporation Method of motion compensation for trans-catheter aortic valve implantation
US20120004533A1 (en) * 2010-05-03 2012-01-05 Slemens Corporation Optimization of multiple candidates in medical device or feature tracking

Patent Citations (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7920911B2 (en) 2000-06-05 2011-04-05 Kabushiki Kaisha Toshiba Interventional MR imaging with detection and display of device position
US6961608B2 (en) 2000-06-05 2005-11-01 Kabushiki Kaisha Toshiba Interventional MR imaging with detection and display of device position
US20020085681A1 (en) 2000-12-28 2002-07-04 Jensen Vernon Thomas Method and apparatus for obtaining and displaying computed tomography images using a fluoroscopy imaging system
WO2003041057A2 (en) 2001-11-08 2003-05-15 The Johns Hopkins University System and method for robot targeting under flouroscopy based on image servoing
US20080033452A1 (en) * 2004-03-03 2008-02-07 Deutsches Krebsforschungszentrum Incremental Real-Time Recording Of Tracked Instruments In Tubular Organ Structures Inside The Human Body
US20050245814A1 (en) * 2004-04-28 2005-11-03 Sunnybrook & Women's College Health Sciences Centre Catheter tracking with phase information
US20100157041A1 (en) * 2007-03-08 2010-06-24 Sync-Rx, Ltd. Automatic stabilization of an image stream of a moving organ
US20100222671A1 (en) * 2007-03-08 2010-09-02 Sync-Rx, Ltd. Identification and presentation of device-to-vessel relative motion
US20100228076A1 (en) * 2007-03-08 2010-09-09 Sync-Rx, Ltd Controlled actuation and deployment of a medical device
US20100290693A1 (en) * 2007-03-08 2010-11-18 Sync-Rx, Ltd. Location-sensitive cursor control and its use for vessel analysis
US20090062641A1 (en) 2007-08-21 2009-03-05 Adrian Barbu Method and system for catheter detection and tracking in a fluoroscopic image sequence
CN101371784A (en) 2007-08-21 2009-02-25 美国西门子医疗解决公司 Method and system for catheter detection and tracking in a fluoroscopic image sequence
US20100249579A1 (en) 2007-12-17 2010-09-30 Starks Daniel R Systems and methods for modeling both unobstructed and obstructed portions of a catheter
CN101221620A (en) 2007-12-20 2008-07-16 北京中星微电子有限公司 Human face tracing method
US20100239148A1 (en) 2009-03-18 2010-09-23 Siemens Corporation Method and System for Automatic Aorta Segmentation
US20110038517A1 (en) 2009-08-17 2011-02-17 Mistretta Charles A System and method for four dimensional angiography and fluoroscopy
CN101661559A (en) 2009-09-16 2010-03-03 中国科学院计算技术研究所 Digital image training and detecting methods
US20110164035A1 (en) 2010-01-07 2011-07-07 Siemens Corporation Method of motion compensation for trans-catheter aortic valve implantation
US20120004533A1 (en) * 2010-05-03 2012-01-05 Slemens Corporation Optimization of multiple candidates in medical device or feature tracking
CN101908153A (en) 2010-08-21 2010-12-08 上海交通大学 Method for estimating head postures in low-resolution image treatment

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
Chinese Office Action dated Sep. 1, 2014 Application No. 201210120135.2.
Collins, et al. "Online selection of discriminative tracking features," 2005, IEEE Trans. on PAMI 27(10), 1631-1643. *
Lin et al. "Object Tracking Using Incremental Fisher Discriminant Analysis," (2004), In: ICPR, vol. 2, pp. 757-760. *
Wang et al. "Hierarchical Guidewire Tracking in Fluoroscopic Sequences", Mar. 27, 2009, Proc. SPIE 7259, Medical Imaging 2009: Image Processing, pp. 1-8. *
Wang et al. "Robust guidewire tracking in fluoroscopy," Computer Vision and Pattern Recognition, Jun. 20-25, 2009. CVPR 2009. IEEE Conference on , pp. 691-698. *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10143438B2 (en) 2015-08-06 2018-12-04 Xiang Zhang System for 3D object modeling and tracking in X-ray imaging
US11823376B2 (en) 2018-05-16 2023-11-21 Benevis Informatics, Llc Systems and methods for review of computer-aided detection of pathology in images

Also Published As

Publication number Publication date
US20120238866A1 (en) 2012-09-20

Similar Documents

Publication Publication Date Title
US9014423B2 (en) Method and system for catheter tracking in fluoroscopic images using adaptive discriminant learning and measurement fusion
US8423121B2 (en) Method and system for guidewire tracking in fluoroscopic image sequences
US8345944B2 (en) System and method for coronary digital subtraction angiography
US8582854B2 (en) Method and system for automatic coronary artery detection
US9155470B2 (en) Method and system for model based fusion on pre-operative computed tomography and intra-operative fluoroscopy using transesophageal echocardiography
US9824302B2 (en) Method and system for model-based fusion of multi-modal volumetric images
Zheng et al. Automatic aorta segmentation and valve landmark detection in C-arm CT: application to aortic valve implantation
US9715637B2 (en) Method and system for automatic aorta segmentation
US9292917B2 (en) Method and system for model-based fusion of computed tomography and non-contrasted C-arm computed tomography
US9002078B2 (en) Method and system for shape-constrained aortic valve landmark detection
US8218845B2 (en) Dynamic pulmonary trunk modeling in computed tomography and magnetic resonance imaging based on the detection of bounding boxes, anatomical landmarks, and ribs of a pulmonary artery
US9999399B2 (en) Method and system for pigtail catheter motion prediction
US20150223773A1 (en) Method and Apparatus for Image Fusion Based Planning of C-Arm Angulation for Structural Heart Disease
US8655041B2 (en) Automatic detection of contrast injection
US10610181B2 (en) Robust calcification tracking in fluoroscopic imaging
US9082158B2 (en) Method and system for real time stent enhancement on live 2D fluoroscopic scene
US8644576B2 (en) Method and system for multi-part left atrium segmentation in C-arm computed tomography volumes using shape constraints
Serradell et al. Simultaneous correspondence and non-rigid 3D reconstruction of the coronary tree from single X-ray images
US9058664B2 (en) 2D-2D fusion for interventional guidance in trans-catheter aortic valve implantation
Mountney et al. Soft tissue tracking for minimally invasive surgery: Learning local deformation online
US9042619B2 (en) Method and system for automatic native and bypass coronary ostia detection in cardiac computed tomography volumes
US9424648B2 (en) Method and system for device detection in 2D medical images
US9524550B2 (en) System and method for coronary digital subtraction angiography
Pinho et al. Assessment and stenting of tracheal stenosis using deformable shape models
US8675914B2 (en) Method and system for needle tracking in fluoroscopic image sequences

Legal Events

Date Code Title Description
AS Assignment

Owner name: SIEMENS CORPORATION, NEW JERSEY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WANG, PENG;ZHENG, YEFENG;FUNKA-LEA, GARETH;AND OTHERS;SIGNING DATES FROM 20120330 TO 20120331;REEL/FRAME:028081/0426

Owner name: SIEMENS AKTIENGESELLSCHAFT, GERMANY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:JOHN, MATTHIAS;BOESE, JAN;SIGNING DATES FROM 20120330 TO 20120404;REEL/FRAME:028081/0398

AS Assignment

Owner name: SIEMENS AKTIENGESELLSCHAFT, GERMANY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SIEMENS CORPORATION;REEL/FRAME:035213/0032

Effective date: 20150309

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: SIEMENS HEALTHCARE GMBH, GERMANY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SIEMENS AKTIENGESELLSCHAFT;REEL/FRAME:039271/0561

Effective date: 20160610

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8

AS Assignment

Owner name: SIEMENS HEALTHINEERS AG, GERMANY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SIEMENS HEALTHCARE GMBH;REEL/FRAME:066088/0256

Effective date: 20231219