US20080009722A1 - Multi-planar reconstruction for ultrasound volume data - Google Patents

Multi-planar reconstruction for ultrasound volume data Download PDF

Info

Publication number
US20080009722A1
US20080009722A1 US11/527,286 US52728606A US2008009722A1 US 20080009722 A1 US20080009722 A1 US 20080009722A1 US 52728606 A US52728606 A US 52728606A US 2008009722 A1 US2008009722 A1 US 2008009722A1
Authority
US
United States
Prior art keywords
planes
data
scanning
ultrasound
orienting
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/527,286
Inventor
Constantine Simopoulos
Lewis J. Thomas
Shaohua Kevin Zhou
Dorin Comaniciu
Bogdan Georgescu
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Siemens Medical Solutions USA Inc
Original Assignee
Siemens Medical Solutions USA Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Siemens Medical Solutions USA Inc filed Critical Siemens Medical Solutions USA Inc
Priority to US11/527,286 priority Critical patent/US20080009722A1/en
Assigned to SIEMENS MEDICAL SOLUTIONS USA, INC. reassignment SIEMENS MEDICAL SOLUTIONS USA, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: THOMAS III, LEWIS J., GEORGESCU, BOGDAN, COMANICIU, DORIN, ZHOU, SHAOHUA KEVIN, SIMOPOULOS, CONSTANTINE
Publication of US20080009722A1 publication Critical patent/US20080009722A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/52Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
    • G01S7/52017Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00 particularly adapted to short-range imaging
    • G01S7/52023Details of receivers
    • G01S7/52036Details of receivers using analysis of echo signal for target characterisation
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/08Detecting organic movements or changes, e.g. tumours, cysts, swellings
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/13Tomography
    • A61B8/14Echo-tomography
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/48Diagnostic techniques
    • A61B8/483Diagnostic techniques involving the acquisition of a 3D volume of data
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/52Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/5215Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data
    • A61B8/523Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data for generating planar views from image data in a user selectable plane not corresponding to the acquisition plane
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S15/00Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
    • G01S15/88Sonar systems specially adapted for specific applications
    • G01S15/89Sonar systems specially adapted for specific applications for mapping or imaging
    • G01S15/8906Short-range imaging systems; Acoustic microscope systems using pulse-echo techniques
    • G01S15/8993Three dimensional imaging systems
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/52Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
    • G01S7/52017Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00 particularly adapted to short-range imaging
    • G01S7/52053Display arrangements
    • G01S7/52057Cathode ray tube displays
    • G01S7/52068Stereoscopic displays; Three-dimensional displays; Pseudo 3D displays
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/52Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
    • G01S7/52017Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00 particularly adapted to short-range imaging
    • G01S7/52085Details related to the ultrasound signal acquisition, e.g. scan sequences
    • G01S7/52087Details related to the ultrasound signal acquisition, e.g. scan sequences using synchronization techniques
    • G01S7/52088Details related to the ultrasound signal acquisition, e.g. scan sequences using synchronization techniques involving retrospective scan line rearrangements
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/08Detecting organic movements or changes, e.g. tumours, cysts, swellings
    • A61B8/0883Detecting organic movements or changes, e.g. tumours, cysts, swellings for diagnosis of the heart
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/08Detecting organic movements or changes, e.g. tumours, cysts, swellings
    • A61B8/0891Detecting organic movements or changes, e.g. tumours, cysts, swellings for diagnosis of blood vessels
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/48Diagnostic techniques
    • A61B8/488Diagnostic techniques involving Doppler signals

Definitions

  • the present embodiments relate to medical diagnostic ultrasound imaging.
  • multi-planar reconstruction for ultrasound volume data is provided.
  • Ultrasound may be used to scan a patient.
  • echocardiography is a commonly used imaging modality to visualize the structure of the heart. Because the echo is often a 2D projection of the 3D human heart, standard views are captured to better visualize the cardiac structures. For example, in the apical four-chamber (A4C) view, all four cavities, namely left and right ventricles, and left and right atria, are present. In the apical two-chamber (A2C) view, only the left ventricle and the left atrium are present.
  • A4C apical four-chamber
  • A2C apical two-chamber
  • Another example is imaging the intracranial structures of a fetus. Three standard planes are acquired with different orientations, not necessarily orthogonal, but fixed with respect to each other, for visualization of the cerebellum, the cistema magna and lateral ventricles.
  • Acquired cardiac or other desired views often deviate from the standard views due to machine properties, the inter-patient variations, or preferences of sonographers.
  • the sonographer manually adjusts imaging parameters of the ultrasound system and transducer position, resulting in variation. For example, the user moves the imaging plane and associated view by moving the transducer relative to the patient. Undesired movement by the patient and/or the sonographer may result in an undesired or non-optimal view for diagnosis.
  • U.S. Published Patent Application No. 2005/0096538 discloses stabilizing the view plane relative to the patient despite some transducer movement. The user positions the plane to a desired location in the patient. Subsequently, the scan plane is varied relative to the transducer to maintain the scan plane in the desired location relative to the patient.
  • Real-time 3D ultrasound is an emerging technique that visualizes a volume region of the patient, such as the human heart in spatial and temporal dimensions. Multiple images may be obtained at a substantially same time.
  • multi-planar reconstruction a volume region is scanned.
  • a plurality of planes such as three planes at substantially right angles to each other, are positioned relative to the volume. Two-dimensional images are generated for each of the planes. However, the views of interest may have different positions relative to each other in the volume. If one plane is aligned to the desired view, other planes may not be aligned.
  • the planes are defined relative to the transducer. Movement of the transducer relative to the patient may result in non-optimal views. Movement of the object of interest, such as fetus, may result in non-optimal views or constant adjustment of the transducer by the sonographer.
  • the preferred embodiments described below include methods, computer readable media and systems for multi-planar reconstruction for ultrasound volume data.
  • a plurality of images is generated corresponding to a plurality of different planes in a volume.
  • the volume scan data is searched by a processor to identify desired views.
  • Multiple standard or predetermined views are generated based on plane positioning within the volume by the processor.
  • Multi-planar reconstruction guided by the processor, allows for real-time imaging of multiple views at a substantially same time.
  • the images corresponding to the identified views are generated independent of the position of the transducer.
  • the planes are positioned in real-time using a pyramid data structure of coarse and fine data sets.
  • a method for multi-planar reconstruction for ultrasound volume data An ultrasound transducer is positioned adjacent, on or within a patient. A volume region is scanned with the ultrasound transducer. A processor determines, from data responsive to the scanning, a first orientation of an object within the volume region while scanning. A multi-planar reconstruction is oriented as a function of the first orientation of the object and independently of a second orientation of the ultrasound transducer relative to the object. Multi-planar reconstruction images of the object are generated from the data while scanning. The images are a function of the orientation of the multi-planar reconstruction.
  • a computer readable storage medium has stored therein data representing instructions executable by a programmed processor for multi-planar reconstruction from ultrasound volume data.
  • the storage medium includes instructions for controlling acquisition by scanning a volume region of a patient; determining, from ultrasound data responsive to the acquisition, locations of features of an object within the volume region represented by the data, the determining being during control of the acquisition by scanning; orienting a plurality of planes within the volume region as a function of the locations of the features, the orienting being independent of an orientation of an ultrasound transducer relative to the object, each of the plurality of planes being different from the other ones of the plurality of planes; and generating images of the object from the data for each of the planes.
  • a method for multi-planar reconstruction from ultrasound volume data Ultrasound data in a first coarse set and a second fine set is obtained.
  • the ultrasound data represents an object in a volume.
  • a processor identifies a plurality of features of the object from the first coarse set of ultrasound data.
  • the processor determines locations of planes for multi-planar reconstruction as a function of the features of the object.
  • the processor refines the locations as a function of the second fine set.
  • FIG. 1 is a block diagram of one embodiment of a medical ultrasound imaging system
  • FIG. 2 is a flow chart diagram of embodiments of methods for multi-planar reconstruction from ultrasound volume data
  • FIG. 3 is a graphical representation of a volume region, object and associated planes of a multi-planar reconstruction in one embodiment
  • FIG. 4 is a graphical representation of directional filters in various embodiments
  • FIG. 5 is a graphical representation of one embodiment of a pyramid data structure
  • FIG. 6 is a graphical representation of one embodiment of an apical four-chamber view.
  • Online or real-time substantially continuous display of different specific anatomical planes may be provided regardless of the orientation of the transducer.
  • a volume is scanned.
  • the data representing the volume is searched for the location of anatomical features associated with planar positions for desired views, such as standard views.
  • a multi-planar reconstruction is provided without the user having to adjust or initially locate a desired view. Multiple views are acquired substantially simultaneously. Since the planes are positioned relative to acquired data and independent of the volume scanning transducer, desired images are generated even where the transducer or imaged object (e.g., heart or fetus) moves. Inexact alignment of the transducer to all of standard planes may be allowed, even for an initial transducer position. Sonographer workflow and acquisition of desired views may be improved.
  • canonical slice(s) or planes such as apical four chamber (A4C) and apical two-chamber (A2C) views, are extracted from the data representing a volume.
  • A4C apical four chamber
  • A2C apical two-chamber
  • These anatomical planes are continuously displayed irrespective of the orientation of the transducer used in the acquisition of the volume ultrasound data. Visualization of the acquired volumetric data may be simplified while scanning, possibly improving workflow.
  • FIG. 1 shows a medical diagnostic imaging system 10 for multi-planar reconstruction from ultrasound volume data.
  • the system 10 is a medical diagnostic ultrasound imaging system, but may be a computer, workstation, database, server, or other system.
  • the system 10 includes a processor 12 , a memory 14 , a display 16 , and a transducer 18 . Additional, different, or fewer components may be provided.
  • the system 10 includes a transmit beamformer, receive beamformer, B-mode detector, Doppler detector, harmonic response detector, contrast agent detector, scan converter, filter, combinations thereof, or other now known or later developed medical diagnostic ultrasound system components.
  • the transducer 18 is a piezoelectric or capacitive device operable to convert between acoustic and electrical energy.
  • the transducer 18 is an array of elements, such as a multi-dimensional or two-dimensional array. Alternatively, the transducer 18 is a wobbler for mechanical scanning in one dimension and electrical scanning in another dimension.
  • the system 10 uses the transducer 18 to scan a volume. Electrical and/or mechanical steering allows transmission and reception along different scan lines in the volume. Any scan pattern may be used.
  • the transmit beam is wide enough for reception along a plurality of scan lines.
  • a plane, collimated or diverging transmit waveform is provided for reception along a plurality, large number, or all scan lines.
  • Ultrasound data representing a volume is provided in response to the scanning.
  • the ultrasound data is beamformed, detected, and/or scan converted.
  • the ultrasound data may be in any format, such as polar coordinate, Cartesian coordinate, a three-dimensional grid, two-dimensional planes in Cartesian coordinate with polar coordinate spacing between planes, or other format.
  • the memory 14 is a buffer, cache, RAM, removable media, hard drive, magnetic, optical, or other now known or later developed memory.
  • the memory 14 is a single device or group of two or more devices.
  • the memory 14 is shown within the system 10 , but may be outside or remote from other components of the system 10 .
  • the memory 14 stores the ultrasound data.
  • the memory 14 stores flow (e.g., velocity, energy or both) and/or B-mode ultrasound data.
  • the medical image data is transferred to the processor 12 from another device.
  • the medical image data is a three-dimensional data set, or a sequence of such sets. For example, a sequence of sets over a portion, one, or more heart cycles of the heart are stored. A plurality of sets may be provided, such as associated with imaging a same person, organ or region from different angles or locations.
  • the ultrasound data bypasses the memory 14 , is temporarily stored in the memory 14 , or is loaded from the memory 14 .
  • Real-time imaging may allow delay of a fraction of seconds, or even seconds, between acquisition of data and imaging.
  • real-time imaging is provided by generating the images substantially simultaneously with the acquisition of the data by scanning. While scanning to acquire a next or subsequent set of data, images are generated for a previous set of data. The imaging occurs during the same imaging session used to acquire the data.
  • the amount of delay between acquisition and imaging for real-time operation may vary, such as a greater delay for initially locating planes of a multi-planar reconstruction with less delay for subsequent imaging.
  • the ultrasound data is stored in the memory 14 from a previous imaging session and used for generating the multi-planar reconstruction without concurrent acquisition.
  • the memory 14 is additionally or alternatively a computer readable storage medium with processing instructions.
  • the memory 14 stores data representing instructions executable by the programmed processor 12 for multi-planar reconstruction for ultrasound volume data.
  • the instructions for implementing the processes, methods and/or techniques discussed herein are provided on computer-readable storage media or memories, such as a cache, buffer, RAM, removable media, hard drive or other computer readable storage media.
  • Computer readable storage media include various types of volatile and nonvolatile storage media. The functions, acts or tasks illustrated in the figures or described herein are executed in response to one or more sets of instructions stored in or on computer readable storage media.
  • the functions, acts or tasks are independent of the particular type of instructions set, storage media, processor or processing strategy and may be performed by software, hardware, integrated circuits, firmware, micro code and the like, operating alone or in combination.
  • processing strategies may include multiprocessing, multitasking, parallel processing and the like.
  • the instructions are stored on a removable media device for reading by local or remote systems.
  • the instructions are stored in a remote location for transfer through a computer network or over telephone lines.
  • the instructions are stored within a given computer, CPU, GPU, or system.
  • the processor 12 is a general processor, digital signal processor, three-dimensional data processor, graphics processing unit, application specific integrated circuit, field programmable gate array, digital circuit, analog circuit, combinations thereof, or other now known or later developed device for processing medical image data.
  • the processor 12 is a single device, a plurality of devices, or a network. For more than one device, parallel or sequential division of processing may be used. Different devices making up the processor 12 may perform different functions, such as a scanning controller and an image generator operating separately.
  • the processor 12 is a control processor or other processor of a medical diagnostic imaging system, such as a medical diagnostic ultrasound imaging system processor.
  • the processor 12 operates pursuant to stored instructions to perform various acts described herein, such as obtaining data, deriving anatomical information, setting an imaging parameter and/or controlling imaging.
  • the processor 12 receives acquired ultrasound data during scanning and determines locations of planes for a multi-planar reconstruction relative to the volume represented by the data.
  • the processor 12 performs or controls other components to perform the methods described herein.
  • the acts of the methods may be implemented by programs and/or a classifier. Any classifier may be applied, such as a model based classifier or a learned classifier or classifier based on machine learning. For learned classifiers, binary or multi-class classifiers may be used, such as Bayesian or neural network classifiers. In one embodiment, a multi-class boosting classifier with a tree and cascade structure is used.
  • the classifier is instructions, a matrix, a learned code, or other software and/or hardware for distinguishing between information in a medical image.
  • Learned feature vectors are used to classify the anatomy.
  • the classifier identifies a canonical view, tissue structure, flow pattern, or combinations thereof from ultrasound data.
  • the classifier may identify cardiac structure associated with a particular view of a heart.
  • the view is a common or standard view (e.g., apical four chamber, apical two chamber, left parasternal, or sub-coastal), but other views may be recognized.
  • the cardiac structure is the heart walls or other structure defining the view or a structure associated with the view. For example, a valve associated with an apical four chamber view is identified.
  • FIG. 2 shows a method for multi-planar reconstruction for ultrasound volume data.
  • the method is implemented by a medical diagnostic imaging system, a review station, a workstation, a computer, a PACS station, a server, combinations thereof, or other device for image processing medical ultrasound data.
  • the system or computer readable media shown in FIG. 1 implements the method, but other systems may be used.
  • the method is implemented in the order shown or a different order. Additional, different, or fewer acts may be performed.
  • acts 22 and/or 34 are optional.
  • scanning is performed in act 26 without controlling the scan.
  • the feedback from act 32 to act 26 may not be provided or may feedback to a different act and/or from a different act.
  • the acts 22 - 34 are performed in real-time, such as during scanning in act 26 .
  • the user may view images of act 32 while scanning in act 26 .
  • the images may be associated with previous performance of acts 22 - 30 in the same imaging session, but with different volume data.
  • acts 22 - 32 are performed for an initial scan.
  • Acts 22 , 26 , 34 and 32 are performed for subsequent scans during the same imaging session.
  • the scan of act 26 may result in images from act 32 in a fraction of a second or longer time period (e.g., seconds) and still be real-time with the scanning.
  • the user is provided with imaging information representing portions of the volume being scanned while scanning.
  • an ultrasound transducer is positioned adjacent, on or within a patient.
  • a volume scanning transducer is positioned, such as a wobbler or multi-dimensional array.
  • the transducer is positioned directly on the skin or acoustically coupled to the skin of the patient.
  • an intraoperative, intercavity, catheter, transesophageal, or other transducer positionable within the patient is used to scan from within the patient.
  • the user may manually position the transducer, such as using a handheld probe or manipulating steering wires.
  • a robotic or mechanical mechanism positions the transducer.
  • act 26 the acquisition of data by scanning a volume region of a patient is controlled.
  • Transmit and receive scanning parameters are set, such as loading a sequence of transmit and receive events to sequentially scan a volume.
  • the transmit and receive beamformers are controlled to acquire ultrasound data representing a volume of the patient adjacent to the transducer.
  • the volume region of the patient is scanned in act 26 .
  • the wobbler or multi-dimensional array generates acoustic energy and receives responsive echoes.
  • a one-dimensional array is manually moved for scanning a volume.
  • the ultrasound data corresponds to a displayed image (e.g., detected and scan converted ultrasound data), beamformed data, detected data, and/or scan converted data.
  • the ultrasound data represents a region of a patient.
  • the region includes tissue, fluid or other structures. Different structures or types of structures react to the ultrasound differently. For example, heart muscle tissue moves, but slowly as compared to fluid. The temporal reaction may result in different velocity or flow data.
  • the shape of a structure or spatial aspect may be reflected in B-mode data.
  • One or more objects, such as the heart, an organ, a vessel, fluid chamber, clot, lesion, muscle, and/or tissue are within the region.
  • the data represents the region.
  • FIG. 3 shows a volume region 40 with an object 42 at least partly within the region 40 .
  • the object 42 may have any orientation within the volume region 40 .
  • the position of planes 44 relative to the object 42 is determined for multi-planar reconstruction.
  • a processor determines from the ultrasound data responsive to the scanning an orientation of the object 42 within the volume region 40 or relative to the transducer in act 28 .
  • the determination is made while scanning in act 26 .
  • the data used for the determination is previously acquired, such as an immediately previous scan, or data presently being acquired.
  • the orientation may be determined using any now known or later developed process.
  • the processor identifies the object without user input.
  • the orientation may be based, in part, on user input.
  • the user indicates the type of organs or object of interest (e.g., selecting cardiology or echocardiography imaging).
  • the location of features is used to determine object orientation.
  • template modeling or matching is used to identify a structure or different structures, such as taught in U.S. Pat. No. 7,092,749, the disclosure of which is incorporated herein by reference.
  • a template is matched to a structure.
  • the template is matched to an overall feature, such as the heart tissue and chambers associated with a standard view.
  • the template may be annotated to identify other features based on the matched view, such as identifying specific chambers or valves.
  • Trained classifiers may be used.
  • Anatomical information is derived from the ultrasound data.
  • the anatomical information is derived from a single set or a sequence of sets.
  • the shape, the position of tissue over time, flow pattern, or other characteristic may indicate anatomical information.
  • Anatomical information includes views, organs, structure, patterns, tissue type, or other information.
  • a feature is any anatomical structure.
  • the anatomical features may be a valve annulus, an apex, chamber, valve, valve flow, or other structure. Features corresponding to a combination of different structures may be used.
  • the locations of one or more features of the object are determined by applying a classifier.
  • a classifier For example, any of the methods disclosed in U.S. Published Patent Application No. ______ (Attorney Docket No. 2006P14951US), the disclosure of which is incorporated herein by reference, is used.
  • the anatomical information is derived by applying a classifier. Any now known or later developed classifier for extracting anatomical information from ultrasound data may be used, such as a single class or binary classifier, collection of different classifiers, cascaded classifiers, hierarchal classifier, multi-class classifier, model based classifier, classifier based on machine learning, or combinations thereof.
  • the classifier is trained from a training data set using a computer.
  • Multi-class classifiers include CART, K-nearest neighbors, neural network (e.g., multi-layer perceptron), mixture models, or others.
  • the AdaBoost.MH algorithm may be used as a multi-class boosting algorithm where no conversion from multi-class to binary is necessary. Error-correcting output code (EOCC) may be used.
  • EOCC Error-correcting output code
  • the classifier is taught to detect objects or information associated with anatomy.
  • the AdaBoost algorithm selectively combines into a strong committee of weak learners based on Haar-like local rectangle filters whose rapid computation is enabled by the use of an integral image.
  • FIG. 4 shows five example filters for locating or highlighting edges.
  • a cascade structure may deal with rare event detection.
  • the FloatBoost a variant of the AdaBoost, may address multiview detection. Multiple objects may be dealt with by training a multi-class classifier with the cascade structure.
  • the classifier learns various feature vectors for distinguishing between classes of features.
  • a probabilistic boosting tree (PBT) which unifies classification, recognition, and clustering into one treatment, may be used.
  • a tree structure may be learned and may offer efficiency in both training and application. Often, in the midst of boosting a multi-class classifier, one class (or several classes) has been completely separated from the remaining ones and further boosting yields no additional improvement in terms of the classification accuracy. For efficient training, a tree structure is trained. To take advantage of this fact, a tree structure is trained by focusing on the remaining classes to improve learning efficiency. Posterior probabilities or known distributions may be computed, such as by correlating anterior probabilities together.
  • a cascade training procedure may be used.
  • a cascade of boosted multi-class strong classifiers may result.
  • the cascade of classifiers provides a unified algorithm able to detect and classify multiple objects while rejecting the background classes.
  • the cascade structure corresponds to a degenerate decision tree.
  • Such a scenario presents an unbalanced nature of data samples.
  • the background class has voluminous samples because all data points not belonging to the object classes belong to the background class. To examine more background examples, only those background examples that pass the early stages of the cascade are used for training a current stage.
  • the trained classifier is applied to the ultrasound data.
  • the ultrasound data is processed to provide the inputs to the classifier.
  • the filters of FIG. 4 are applied to the ultrasound data.
  • a matrix or vector representing the outputs of the filters is input to the classifier.
  • the classifier identifies features based on the inputs.
  • the anatomical information is encoded based on a template.
  • the image is searched based on the template information, localizing the chambers or other structure.
  • a search from the left-top corner to the right-bottom corner is performed by changing the width, height, and angle of a template box.
  • the search is performed in a pyramid structure, with a coarse search on a lower resolution or decimated image and a refined search on a higher resolution image based on the results of the coarse search.
  • This exhaustive search approach may yield multiple results of detections and classifications, especially around the correct view location.
  • the search identifies relevant features.
  • FIG. 3 shows a plurality of features 46 .
  • FIG. 6 shows an A4C view with three identified features.
  • a multi-planar reconstruction is oriented as a function of the orientation of the object.
  • the orientation of the object may be represented by two or more features. By identifying the location of features of the object, the position of the object relative to the volume region 40 and/or the transducer is determined.
  • Two or more planes 44 are oriented within the volume region 40 as a function of the locations of the features. Each of the planes is different, such as being defined by different features.
  • a plane 44 may be oriented based on identification of a single feature, such as a view. Alternatively, the plane 44 is oriented based on the identification of a linear feature and a point feature or three or more point features. Any combination of features defining a plane 44 may be used. For example, an apex, four chambers, and a valve annulus define an apical four-chamber view. The apex, two chambers and a valve annulus define an apical two-chamber view in the same volume region 40 .
  • the planes are oriented independent of an orientation of the ultrasound transducer relative to the object 42 . Since the search for features is performed in three-dimensional space, the features may be identified without consideration of the orientation of the transducer.
  • the transducer does not need to be moved to define a view or establish a plane.
  • the planes are found based on the ultrasound data representing the volume. Orienting is performed without fixing a view relative to the transducer and by fixing the view relative to the object. Since the orientation is performed during the control of the acquisition by scanning, the user may, but is not required to, precisely position the transducer.
  • the features and corresponding plane orientations identify a standard or predetermined view.
  • One or more of the planes 44 of the reconstruction are positioned relative to the object 42 such that the planes correspond to the standard and/or predetermined view.
  • Standard views may be standard for the medical community or standard for an institution.
  • the object is a heart
  • the planes are positioned to provide an apical two chamber, apical four chamber, parasternal long axis, and/or parasternal short axis view.
  • Predetermined views include non-standard views, such as a pre-defined view for clinical testing.
  • One, two, or more planes are oriented to provide different views.
  • three, four or more planes are oriented, such as for echocardiography.
  • three planes e.g., two longitudinal views at different angles of rotation about a longitudinal axis and one cross-sectional view
  • One or more planes may be for non-standard and/or non-predetermined views. Each plane is independently oriented based on features. Alternatively or additionally, the orientation of one plane may be used to determine an orientation of another plane.
  • multi-planar reconstruction images of the object are generated from the ultrasound data.
  • the planes define the data to be used for imaging. Data associated with locations intersecting each plane or adjacent to each plane is used to generate a two-dimensional image. Data may be interpolated to provide spatial alignment to the plane, or a nearest neighbor selection may be used.
  • the resulting images are generated as a function of the orientation of the multi-planar reconstruction and provide the desired view.
  • the images represent different planes 44 through the volume region 40 .
  • specific views are generated. All or a sub-set of the specific views are generated. Where planes corresponding to the views are identified, the views may be provided. For example, all the available standard or predetermined views in ultrasound data representing a region are provided. The images for each view may be labeled (e.g., A4C) and/or annotated (e.g., valve highlighted). Fewer than all available views may be provided, such as displaying no more than three views and having a priority list of views.
  • the images are generated during the control of the acquisition by scanning. Real-time generation allows a sonographer to verify the desired information or images are obtained.
  • FIG. 2 shows a feedback for repeating the determining act 28 , orienting act 30 , and generating act 32 a plurality of times while scanning (act 26 ) in a same imaging session.
  • the generating of act 32 may be performed more frequently, such as applying previously determined plane positions to subsequent data sets.
  • the determining and orienting acts 28 , 30 may be performed less frequently.
  • the determining and orienting acts 28 , 30 are performed for each set of ultrasound data representing the volume region.
  • the repetition of the determining and orienting acts 28 , 30 may be periodic. Every set, second set, third set or other number of sets of data acquired, the acts are triggered.
  • the trigger may be based on timing, a clock, or a count.
  • Other trigger events may be used, such as heart cycle events, detected transducer motion, detected object motion, or failure of tracking.
  • the features or views are tracked between sets of data and the plane positions are refined based on the tracking.
  • the anatomical view planes of the multi-planar reconstruction are tracked as a function of time. For example, speckle or feature tracking (e.g., correlation or minimum sum of absolute differences) of the features is performed as a function of time. As another example, velocity estimates are used to indicate displacement. Other tracking techniques may be used.
  • the amount of displacement, direction of displacement, and/or rotation of the features or two-dimensional image region defining a plane is determined.
  • the plane position is adjusted based on the tracking. Images are generated in act 32 using the adjusted planes.
  • One example embodiment for determining an orientation of planes in a multi-planar reconstruction uses a pyramid of ultrasound data sets. More rapid determination may be provided.
  • the ultrasound data is used to create two or more sets of data with different resolution (see FIG. 5 for a graphical example). For example, a one set of data has fine resolution, such as the scan resolution, and another set of data has a coarse resolution, such as the fine set decimated by 1 ⁇ 2 in each dimension.
  • the sets represent the same object in the same volume. Three or more sets may be used.
  • the sets are in any format, such as Cartesian or polar coordinate.
  • the ultrasound data is acquired in an acoustic (e.g., polar) coordinate format, and the Cartesian or display space is populated in real time with only visible surfaces or selected planes.
  • a scan converter, processor, or graphics processing unit real-time conversion from the acoustic space to the Cartesian or display space is feasible.
  • the ultrasound data is processed in the Cartesian space (e.g., 3D grid) to orient the multi-planar reconstruction.
  • a hierarchical coarse-to-fine strategy is implemented.
  • the data is processed on the highest level, such as scanning for feature points and/or planes.
  • the subsequent levels are processed for further details.
  • the pyramid processing may improve the processing speed.
  • a single set of ultrasound data is used without the data pyramid.
  • the locations of multiple planes are detected.
  • the 3D volumetric data is stored in the spherical coordinates.
  • the relationship between the Cartesian coordinates (x, y, s) and the spherical coordinates (r, ⁇ , ⁇ ) is as follows:
  • an estimate of the multiple planes is performed.
  • the data is processed in a hierarchical manner to detect reliable feature points, such as an apex, valve annulus points, and/or chambers (e.g., the chamber of the left ventricle).
  • reliable feature points such as an apex, valve annulus points, and/or chambers (e.g., the chamber of the left ventricle).
  • the features provide a rough estimate of the positions of the multiple planes.
  • the whole heart or structure may be aligned to a canonical model.
  • Object features are identified.
  • a processor identifies a plurality of features of the object from the first coarse set of ultrasound data. Larger features may be used for determining the locations in the coarse set. The features indicate the likely location of smaller features. The smaller features are then identified in the fine set of the volume pyramid based on the larger features. Alternatively, the smaller features are identified regardless of the larger features. The location of the larger features may be refined using the fine set with or without identifying other features in the fine set.
  • the processor determines locations of planes for multi-planar reconstruction as a function of the features of the object. As discussed above for act 30 , the position of planes of the multi-planar reconstruction is located as a function of the object features. Images may then be generated.
  • One embodiment is for multi-plane detection and tracking in real-time 3D echocardiography.
  • the location for the multiple planes is initially determined without required user adjustment of an imaging or scanning plane. Once the positions are determined, the planes are tracked for subsequent imaging using a different approach.
  • detection for refinement may be programmed.
  • a learning approach to detection is used.
  • An annotated database is obtained. Training positive and negative examples are extracted from the database.
  • a refined estimate of the positions of the multiple planes is determined from the set of data with a finer resolution.
  • the estimates of each plane positioning are refined. Any now known or later developed refinement may be used.
  • a designed object detector that discriminates the subtle difference in the appearance is trained, or a regressor is learned to infer the estimate.
  • projection pursuit is used.
  • Points like the apex and valve annulus present distinct structures that may be well described by specified features.
  • the heart chambers have semantic constraints that could be learned by higher-level classifiers.
  • a hierarchical representation, modeling, and detection of the heart structures may be used.
  • plane-specific detectors are learned.
  • One detector for one particular plane may be learned.
  • an A4C detector and an A2C detector are learned separately.
  • a joint detector for multiple planes that are treated as an entity is learned.
  • the detector is a binary classifier F(I):
  • the binary classifier known as the AdaBoosting algorithm, probabilistic boosting tree (PDT) or other classifiers may be used.
  • the boosting algorithm is used to select Haar features, which are amenable to rapid evaluation.
  • the Haar features are combined to form a committee classifier.
  • An image-based multi-class classifier to further differentiate multiple planes may be used in addition.
  • ⁇ ⁇ arg ⁇ ⁇ max ⁇ ⁇ ⁇ p ⁇ ( V ⁇ ( ⁇ ) )
  • the posterior probability p(V(a)) is replaced by a pursuit cost function.
  • a pursuit cost function For example, it can be based on projected histogram or other measures.
  • Regression may be used for refinement with the fine set of ultrasound data.
  • One practical consideration is the searching speed especially when confronting the three-dimensional parameter space. To accelerate the speed, a function:
  • is the difference between the current parameter ⁇ and the ground truth ⁇ acute over ( ⁇ ) ⁇ , as represented by:
  • the locations of the planes are refined.
  • An image-based regression task is used to find the target function g(I) that minimizes the cost function:
  • L(g(l n ), ⁇ n ) is the loss function and R(g) is a regularization term.
  • the planes are tracked. In real-time acquisition, after locking the multiple planes, the planes are tracked. Any tracking may be used. In an alternative embodiment, the detector is applied for every scan or ultrasound data representing the volume region.
  • Another approach is to use multiple hypotheses tracking.
  • hypotheses of a are generated and verified by the filtering probability p( ⁇ 1
  • u t is a noise component. If the classifier p(I) is used as a likelihood measurement q(V t
  • ⁇ t ) p(V t ( ⁇ t )), the posterior filtering probability p( ⁇ t
  • the Kalman filter may provide an exact solution. Often, the above time series system is not linear and the distributions are non-Gaussian.
  • the sequential Monte Carlo (SMC) algorithm may be used to derive an approximate solution, allowing real-time tracking.

Abstract

During scanning or in real-time with acquisition of ultrasound data, a plurality of images is generated corresponding to a plurality of different planes in a volume. The volume scan data is searched by a processor to identify desired views. Multiple standard or predetermined views are generated based on plane positioning within the volume by the processor. Multi-planar reconstruction, guided by the processor, allows for real-time imaging of multiple views at a substantially same time. The images corresponding to the identified views are generated independent of the position of the transducer. The planes may be positioned in real-time using a pyramid data structure of coarse and fine data sets.

Description

    RELATED APPLICATIONS
  • The present patent document claims the benefit of the filing date under 35 U.S.C. §119(e) of Provisional U.S. Patent Application Ser. No. 60/747,024, filed May 11, 2006, which is hereby incorporated by reference.
  • BACKGROUND
  • The present embodiments relate to medical diagnostic ultrasound imaging. In particular, multi-planar reconstruction for ultrasound volume data is provided.
  • Ultrasound may be used to scan a patient. For example, echocardiography is a commonly used imaging modality to visualize the structure of the heart. Because the echo is often a 2D projection of the 3D human heart, standard views are captured to better visualize the cardiac structures. For example, in the apical four-chamber (A4C) view, all four cavities, namely left and right ventricles, and left and right atria, are present. In the apical two-chamber (A2C) view, only the left ventricle and the left atrium are present. Another example is imaging the intracranial structures of a fetus. Three standard planes are acquired with different orientations, not necessarily orthogonal, but fixed with respect to each other, for visualization of the cerebellum, the cistema magna and lateral ventricles.
  • Acquired cardiac or other desired views often deviate from the standard views due to machine properties, the inter-patient variations, or preferences of sonographers. The sonographer manually adjusts imaging parameters of the ultrasound system and transducer position, resulting in variation. For example, the user moves the imaging plane and associated view by moving the transducer relative to the patient. Undesired movement by the patient and/or the sonographer may result in an undesired or non-optimal view for diagnosis. U.S. Published Patent Application No. 2005/0096538 discloses stabilizing the view plane relative to the patient despite some transducer movement. The user positions the plane to a desired location in the patient. Subsequently, the scan plane is varied relative to the transducer to maintain the scan plane in the desired location relative to the patient.
  • Real-time 3D ultrasound, such as in echocardiography, is an emerging technique that visualizes a volume region of the patient, such as the human heart in spatial and temporal dimensions. Multiple images may be obtained at a substantially same time. In multi-planar reconstruction, a volume region is scanned. Rather than or in addition to three-dimensional rendering, a plurality of planes, such as three planes at substantially right angles to each other, are positioned relative to the volume. Two-dimensional images are generated for each of the planes. However, the views of interest may have different positions relative to each other in the volume. If one plane is aligned to the desired view, other planes may not be aligned. For real-time scanning, the planes are defined relative to the transducer. Movement of the transducer relative to the patient may result in non-optimal views. Movement of the object of interest, such as fetus, may result in non-optimal views or constant adjustment of the transducer by the sonographer.
  • BRIEF SUMMARY
  • By way of introduction, the preferred embodiments described below include methods, computer readable media and systems for multi-planar reconstruction for ultrasound volume data. During scanning or in real-time with acquisition of ultrasound data, a plurality of images is generated corresponding to a plurality of different planes in a volume. The volume scan data is searched by a processor to identify desired views. Multiple standard or predetermined views are generated based on plane positioning within the volume by the processor. Multi-planar reconstruction, guided by the processor, allows for real-time imaging of multiple views at a substantially same time. The images corresponding to the identified views are generated independent of the position of the transducer. In a same or other embodiment, the planes are positioned in real-time using a pyramid data structure of coarse and fine data sets.
  • In a first aspect, a method is provided for multi-planar reconstruction for ultrasound volume data. An ultrasound transducer is positioned adjacent, on or within a patient. A volume region is scanned with the ultrasound transducer. A processor determines, from data responsive to the scanning, a first orientation of an object within the volume region while scanning. A multi-planar reconstruction is oriented as a function of the first orientation of the object and independently of a second orientation of the ultrasound transducer relative to the object. Multi-planar reconstruction images of the object are generated from the data while scanning. The images are a function of the orientation of the multi-planar reconstruction.
  • In a second aspect, a computer readable storage medium has stored therein data representing instructions executable by a programmed processor for multi-planar reconstruction from ultrasound volume data. The storage medium includes instructions for controlling acquisition by scanning a volume region of a patient; determining, from ultrasound data responsive to the acquisition, locations of features of an object within the volume region represented by the data, the determining being during control of the acquisition by scanning; orienting a plurality of planes within the volume region as a function of the locations of the features, the orienting being independent of an orientation of an ultrasound transducer relative to the object, each of the plurality of planes being different from the other ones of the plurality of planes; and generating images of the object from the data for each of the planes.
  • In a third aspect, a method is provided for multi-planar reconstruction from ultrasound volume data. Ultrasound data in a first coarse set and a second fine set is obtained. The ultrasound data represents an object in a volume. A processor identifies a plurality of features of the object from the first coarse set of ultrasound data. The processor determines locations of planes for multi-planar reconstruction as a function of the features of the object. The processor refines the locations as a function of the second fine set.
  • The present invention is defined by the following claims, and nothing in this section should be taken as a limitation on those claims. Further aspects and advantages of the invention are discussed below in conjunction with the preferred embodiments and may be later claimed independently or in combination.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The components and the figures are not necessarily to scale, emphasis instead being placed upon illustrating the principles of the invention. Moreover, in the figures, like reference numerals designate corresponding parts throughout the different views.
  • FIG. 1 is a block diagram of one embodiment of a medical ultrasound imaging system;
  • FIG. 2 is a flow chart diagram of embodiments of methods for multi-planar reconstruction from ultrasound volume data;
  • FIG. 3 is a graphical representation of a volume region, object and associated planes of a multi-planar reconstruction in one embodiment;
  • FIG. 4 is a graphical representation of directional filters in various embodiments;
  • FIG. 5 is a graphical representation of one embodiment of a pyramid data structure; and
  • FIG. 6 is a graphical representation of one embodiment of an apical four-chamber view.
  • DETAILED DESCRIPTION OF THE DRAWINGS AND PRESENTLY PREFERRED EMBODIMENTS
  • Online or real-time substantially continuous display of different specific anatomical planes may be provided regardless of the orientation of the transducer. A volume is scanned. The data representing the volume is searched for the location of anatomical features associated with planar positions for desired views, such as standard views. A multi-planar reconstruction is provided without the user having to adjust or initially locate a desired view. Multiple views are acquired substantially simultaneously. Since the planes are positioned relative to acquired data and independent of the volume scanning transducer, desired images are generated even where the transducer or imaged object (e.g., heart or fetus) moves. Inexact alignment of the transducer to all of standard planes may be allowed, even for an initial transducer position. Sonographer workflow and acquisition of desired views may be improved.
  • In an echocardiography example, canonical slice(s) or planes, such as apical four chamber (A4C) and apical two-chamber (A2C) views, are extracted from the data representing a volume. These anatomical planes are continuously displayed irrespective of the orientation of the transducer used in the acquisition of the volume ultrasound data. Visualization of the acquired volumetric data may be simplified while scanning, possibly improving workflow.
  • FIG. 1 shows a medical diagnostic imaging system 10 for multi-planar reconstruction from ultrasound volume data. The system 10 is a medical diagnostic ultrasound imaging system, but may be a computer, workstation, database, server, or other system.
  • The system 10 includes a processor 12, a memory 14, a display 16, and a transducer 18. Additional, different, or fewer components may be provided. For example, the system 10 includes a transmit beamformer, receive beamformer, B-mode detector, Doppler detector, harmonic response detector, contrast agent detector, scan converter, filter, combinations thereof, or other now known or later developed medical diagnostic ultrasound system components.
  • The transducer 18 is a piezoelectric or capacitive device operable to convert between acoustic and electrical energy. The transducer 18 is an array of elements, such as a multi-dimensional or two-dimensional array. Alternatively, the transducer 18 is a wobbler for mechanical scanning in one dimension and electrical scanning in another dimension.
  • The system 10 uses the transducer 18 to scan a volume. Electrical and/or mechanical steering allows transmission and reception along different scan lines in the volume. Any scan pattern may be used. In one embodiment, the transmit beam is wide enough for reception along a plurality of scan lines. In another embodiment, a plane, collimated or diverging transmit waveform is provided for reception along a plurality, large number, or all scan lines.
  • Ultrasound data representing a volume is provided in response to the scanning. The ultrasound data is beamformed, detected, and/or scan converted. The ultrasound data may be in any format, such as polar coordinate, Cartesian coordinate, a three-dimensional grid, two-dimensional planes in Cartesian coordinate with polar coordinate spacing between planes, or other format.
  • The memory 14 is a buffer, cache, RAM, removable media, hard drive, magnetic, optical, or other now known or later developed memory. The memory 14 is a single device or group of two or more devices. The memory 14 is shown within the system 10, but may be outside or remote from other components of the system 10.
  • The memory 14 stores the ultrasound data. For example, the memory 14 stores flow (e.g., velocity, energy or both) and/or B-mode ultrasound data. Alternatively, the medical image data is transferred to the processor 12 from another device. The medical image data is a three-dimensional data set, or a sequence of such sets. For example, a sequence of sets over a portion, one, or more heart cycles of the heart are stored. A plurality of sets may be provided, such as associated with imaging a same person, organ or region from different angles or locations.
  • For real-time imaging, the ultrasound data bypasses the memory 14, is temporarily stored in the memory 14, or is loaded from the memory 14. Real-time imaging may allow delay of a fraction of seconds, or even seconds, between acquisition of data and imaging. For example, real-time imaging is provided by generating the images substantially simultaneously with the acquisition of the data by scanning. While scanning to acquire a next or subsequent set of data, images are generated for a previous set of data. The imaging occurs during the same imaging session used to acquire the data. The amount of delay between acquisition and imaging for real-time operation may vary, such as a greater delay for initially locating planes of a multi-planar reconstruction with less delay for subsequent imaging. In alternative embodiments, the ultrasound data is stored in the memory 14 from a previous imaging session and used for generating the multi-planar reconstruction without concurrent acquisition.
  • The memory 14 is additionally or alternatively a computer readable storage medium with processing instructions. The memory 14 stores data representing instructions executable by the programmed processor 12 for multi-planar reconstruction for ultrasound volume data. The instructions for implementing the processes, methods and/or techniques discussed herein are provided on computer-readable storage media or memories, such as a cache, buffer, RAM, removable media, hard drive or other computer readable storage media. Computer readable storage media include various types of volatile and nonvolatile storage media. The functions, acts or tasks illustrated in the figures or described herein are executed in response to one or more sets of instructions stored in or on computer readable storage media. The functions, acts or tasks are independent of the particular type of instructions set, storage media, processor or processing strategy and may be performed by software, hardware, integrated circuits, firmware, micro code and the like, operating alone or in combination. Likewise, processing strategies may include multiprocessing, multitasking, parallel processing and the like. In one embodiment, the instructions are stored on a removable media device for reading by local or remote systems. In other embodiments, the instructions are stored in a remote location for transfer through a computer network or over telephone lines. In yet other embodiments, the instructions are stored within a given computer, CPU, GPU, or system.
  • The processor 12 is a general processor, digital signal processor, three-dimensional data processor, graphics processing unit, application specific integrated circuit, field programmable gate array, digital circuit, analog circuit, combinations thereof, or other now known or later developed device for processing medical image data. The processor 12 is a single device, a plurality of devices, or a network. For more than one device, parallel or sequential division of processing may be used. Different devices making up the processor 12 may perform different functions, such as a scanning controller and an image generator operating separately. In one embodiment, the processor 12 is a control processor or other processor of a medical diagnostic imaging system, such as a medical diagnostic ultrasound imaging system processor. The processor 12 operates pursuant to stored instructions to perform various acts described herein, such as obtaining data, deriving anatomical information, setting an imaging parameter and/or controlling imaging.
  • In one embodiment, the processor 12 receives acquired ultrasound data during scanning and determines locations of planes for a multi-planar reconstruction relative to the volume represented by the data. The processor 12 performs or controls other components to perform the methods described herein. The acts of the methods may be implemented by programs and/or a classifier. Any classifier may be applied, such as a model based classifier or a learned classifier or classifier based on machine learning. For learned classifiers, binary or multi-class classifiers may be used, such as Bayesian or neural network classifiers. In one embodiment, a multi-class boosting classifier with a tree and cascade structure is used. The classifier is instructions, a matrix, a learned code, or other software and/or hardware for distinguishing between information in a medical image. Learned feature vectors are used to classify the anatomy. For example, the classifier identifies a canonical view, tissue structure, flow pattern, or combinations thereof from ultrasound data. In cardiac imaging, the classifier may identify cardiac structure associated with a particular view of a heart. The view is a common or standard view (e.g., apical four chamber, apical two chamber, left parasternal, or sub-coastal), but other views may be recognized. The cardiac structure is the heart walls or other structure defining the view or a structure associated with the view. For example, a valve associated with an apical four chamber view is identified.
  • FIG. 2 shows a method for multi-planar reconstruction for ultrasound volume data. The method is implemented by a medical diagnostic imaging system, a review station, a workstation, a computer, a PACS station, a server, combinations thereof, or other device for image processing medical ultrasound data. For example, the system or computer readable media shown in FIG. 1 implements the method, but other systems may be used. The method is implemented in the order shown or a different order. Additional, different, or fewer acts may be performed. For example, acts 22 and/or 34 are optional. As another example, scanning is performed in act 26 without controlling the scan. The feedback from act 32 to act 26 may not be provided or may feedback to a different act and/or from a different act.
  • The acts 22-34 are performed in real-time, such as during scanning in act 26. The user may view images of act 32 while scanning in act 26. The images may be associated with previous performance of acts 22-30 in the same imaging session, but with different volume data. For example, acts 22-32 are performed for an initial scan. Acts 22, 26, 34 and 32 are performed for subsequent scans during the same imaging session. The scan of act 26 may result in images from act 32 in a fraction of a second or longer time period (e.g., seconds) and still be real-time with the scanning. The user is provided with imaging information representing portions of the volume being scanned while scanning.
  • In act 22, an ultrasound transducer is positioned adjacent, on or within a patient. A volume scanning transducer is positioned, such as a wobbler or multi-dimensional array. For adjacent or on a patient, the transducer is positioned directly on the skin or acoustically coupled to the skin of the patient. For within the patient, an intraoperative, intercavity, catheter, transesophageal, or other transducer positionable within the patient is used to scan from within the patient.
  • The user may manually position the transducer, such as using a handheld probe or manipulating steering wires. Alternatively, a robotic or mechanical mechanism positions the transducer.
  • In act 26, the acquisition of data by scanning a volume region of a patient is controlled. Transmit and receive scanning parameters are set, such as loading a sequence of transmit and receive events to sequentially scan a volume. The transmit and receive beamformers are controlled to acquire ultrasound data representing a volume of the patient adjacent to the transducer.
  • In response to the control, the volume region of the patient is scanned in act 26. The wobbler or multi-dimensional array generates acoustic energy and receives responsive echoes. In alternative embodiments, a one-dimensional array is manually moved for scanning a volume.
  • One or more sets of data are obtained. The ultrasound data corresponds to a displayed image (e.g., detected and scan converted ultrasound data), beamformed data, detected data, and/or scan converted data. The ultrasound data represents a region of a patient. The region includes tissue, fluid or other structures. Different structures or types of structures react to the ultrasound differently. For example, heart muscle tissue moves, but slowly as compared to fluid. The temporal reaction may result in different velocity or flow data. The shape of a structure or spatial aspect may be reflected in B-mode data. One or more objects, such as the heart, an organ, a vessel, fluid chamber, clot, lesion, muscle, and/or tissue are within the region. The data represents the region.
  • For example, FIG. 3 shows a volume region 40 with an object 42 at least partly within the region 40. The object 42 may have any orientation within the volume region 40. The position of planes 44 relative to the object 42 is determined for multi-planar reconstruction.
  • Referring to FIGS. 2 and 3, a processor determines from the ultrasound data responsive to the scanning an orientation of the object 42 within the volume region 40 or relative to the transducer in act 28. The determination is made while scanning in act 26. The data used for the determination is previously acquired, such as an immediately previous scan, or data presently being acquired.
  • The orientation may be determined using any now known or later developed process. The processor identifies the object without user input. Alternatively, the orientation may be based, in part, on user input. For example, the user indicates the type of organs or object of interest (e.g., selecting cardiology or echocardiography imaging).
  • The location of features is used to determine object orientation. For example, template modeling or matching is used to identify a structure or different structures, such as taught in U.S. Pat. No. 7,092,749, the disclosure of which is incorporated herein by reference. A template is matched to a structure. In one embodiment, the template is matched to an overall feature, such as the heart tissue and chambers associated with a standard view. The template may be annotated to identify other features based on the matched view, such as identifying specific chambers or valves.
  • Trained classifiers may be used. Anatomical information is derived from the ultrasound data. The anatomical information is derived from a single set or a sequence of sets. For example, the shape, the position of tissue over time, flow pattern, or other characteristic may indicate anatomical information. Anatomical information includes views, organs, structure, patterns, tissue type, or other information. A feature is any anatomical structure. For cardiac imaging, the anatomical features may be a valve annulus, an apex, chamber, valve, valve flow, or other structure. Features corresponding to a combination of different structures may be used.
  • In one embodiment, the locations of one or more features of the object are determined by applying a classifier. For example, any of the methods disclosed in U.S. Published Patent Application No. ______ (Attorney Docket No. 2006P14951US), the disclosure of which is incorporated herein by reference, is used. The anatomical information is derived by applying a classifier. Any now known or later developed classifier for extracting anatomical information from ultrasound data may be used, such as a single class or binary classifier, collection of different classifiers, cascaded classifiers, hierarchal classifier, multi-class classifier, model based classifier, classifier based on machine learning, or combinations thereof. The classifier is trained from a training data set using a computer. Multi-class classifiers include CART, K-nearest neighbors, neural network (e.g., multi-layer perceptron), mixture models, or others. The AdaBoost.MH algorithm may be used as a multi-class boosting algorithm where no conversion from multi-class to binary is necessary. Error-correcting output code (EOCC) may be used.
  • For learning-based approaches, the classifier is taught to detect objects or information associated with anatomy. For example, the AdaBoost algorithm selectively combines into a strong committee of weak learners based on Haar-like local rectangle filters whose rapid computation is enabled by the use of an integral image. FIG. 4 shows five example filters for locating or highlighting edges. A cascade structure may deal with rare event detection. The FloatBoost, a variant of the AdaBoost, may address multiview detection. Multiple objects may be dealt with by training a multi-class classifier with the cascade structure. The classifier learns various feature vectors for distinguishing between classes of features. A probabilistic boosting tree (PBT), which unifies classification, recognition, and clustering into one treatment, may be used.
  • A tree structure may be learned and may offer efficiency in both training and application. Often, in the midst of boosting a multi-class classifier, one class (or several classes) has been completely separated from the remaining ones and further boosting yields no additional improvement in terms of the classification accuracy. For efficient training, a tree structure is trained. To take advantage of this fact, a tree structure is trained by focusing on the remaining classes to improve learning efficiency. Posterior probabilities or known distributions may be computed, such as by correlating anterior probabilities together.
  • To handle the background classes with many examples, a cascade training procedure may be used. A cascade of boosted multi-class strong classifiers may result. The cascade of classifiers provides a unified algorithm able to detect and classify multiple objects while rejecting the background classes. The cascade structure corresponds to a degenerate decision tree. Such a scenario presents an unbalanced nature of data samples. The background class has voluminous samples because all data points not belonging to the object classes belong to the background class. To examine more background examples, only those background examples that pass the early stages of the cascade are used for training a current stage.
  • The trained classifier is applied to the ultrasound data. The ultrasound data is processed to provide the inputs to the classifier. For example, the filters of FIG. 4 are applied to the ultrasound data. For each spatial location, a matrix or vector representing the outputs of the filters is input to the classifier. The classifier identifies features based on the inputs.
  • The anatomical information is encoded based on a template. The image is searched based on the template information, localizing the chambers or other structure. Given a sequence of medical images, a search from the left-top corner to the right-bottom corner is performed by changing the width, height, and angle of a template box. The search is performed in a pyramid structure, with a coarse search on a lower resolution or decimated image and a refined search on a higher resolution image based on the results of the coarse search. This exhaustive search approach may yield multiple results of detections and classifications, especially around the correct view location. The search identifies relevant features. FIG. 3 shows a plurality of features 46. FIG. 6 shows an A4C view with three identified features.
  • In act 30, a multi-planar reconstruction is oriented as a function of the orientation of the object. The orientation of the object may be represented by two or more features. By identifying the location of features of the object, the position of the object relative to the volume region 40 and/or the transducer is determined.
  • Two or more planes 44 are oriented within the volume region 40 as a function of the locations of the features. Each of the planes is different, such as being defined by different features. A plane 44 may be oriented based on identification of a single feature, such as a view. Alternatively, the plane 44 is oriented based on the identification of a linear feature and a point feature or three or more point features. Any combination of features defining a plane 44 may be used. For example, an apex, four chambers, and a valve annulus define an apical four-chamber view. The apex, two chambers and a valve annulus define an apical two-chamber view in the same volume region 40.
  • The planes are oriented independent of an orientation of the ultrasound transducer relative to the object 42. Since the search for features is performed in three-dimensional space, the features may be identified without consideration of the orientation of the transducer. The transducer does not need to be moved to define a view or establish a plane. The planes are found based on the ultrasound data representing the volume. Orienting is performed without fixing a view relative to the transducer and by fixing the view relative to the object. Since the orientation is performed during the control of the acquisition by scanning, the user may, but is not required to, precisely position the transducer.
  • The features and corresponding plane orientations identify a standard or predetermined view. One or more of the planes 44 of the reconstruction are positioned relative to the object 42 such that the planes correspond to the standard and/or predetermined view. Standard views may be standard for the medical community or standard for an institution. For example in cardiac imaging, the object is a heart, and the planes are positioned to provide an apical two chamber, apical four chamber, parasternal long axis, and/or parasternal short axis view. Predetermined views include non-standard views, such as a pre-defined view for clinical testing.
  • One, two, or more planes are oriented to provide different views. In one embodiment, three, four or more planes are oriented, such as for echocardiography. In another embodiment, three planes (e.g., two longitudinal views at different angles of rotation about a longitudinal axis and one cross-sectional view) are oriented relative to three dimensions along a flow direction of a vessel. One or more planes may be for non-standard and/or non-predetermined views. Each plane is independently oriented based on features. Alternatively or additionally, the orientation of one plane may be used to determine an orientation of another plane.
  • In act 32, multi-planar reconstruction images of the object are generated from the ultrasound data. The planes define the data to be used for imaging. Data associated with locations intersecting each plane or adjacent to each plane is used to generate a two-dimensional image. Data may be interpolated to provide spatial alignment to the plane, or a nearest neighbor selection may be used. The resulting images are generated as a function of the orientation of the multi-planar reconstruction and provide the desired view. The images represent different planes 44 through the volume region 40.
  • In one embodiment, specific views are generated. All or a sub-set of the specific views are generated. Where planes corresponding to the views are identified, the views may be provided. For example, all the available standard or predetermined views in ultrasound data representing a region are provided. The images for each view may be labeled (e.g., A4C) and/or annotated (e.g., valve highlighted). Fewer than all available views may be provided, such as displaying no more than three views and having a priority list of views.
  • The images are generated during the control of the acquisition by scanning. Real-time generation allows a sonographer to verify the desired information or images are obtained.
  • FIG. 2 shows a feedback for repeating the determining act 28, orienting act 30, and generating act 32 a plurality of times while scanning (act 26) in a same imaging session. The generating of act 32 may be performed more frequently, such as applying previously determined plane positions to subsequent data sets. The determining and orienting acts 28, 30 may be performed less frequently. Alternatively, the determining and orienting acts 28, 30 are performed for each set of ultrasound data representing the volume region.
  • The repetition of the determining and orienting acts 28, 30 may be periodic. Every set, second set, third set or other number of sets of data acquired, the acts are triggered. The trigger may be based on timing, a clock, or a count. Other trigger events may be used, such as heart cycle events, detected transducer motion, detected object motion, or failure of tracking.
  • As an alternative or in addition to repeating the determining and orienting acts 28, 30, the features or views are tracked between sets of data and the plane positions are refined based on the tracking. The anatomical view planes of the multi-planar reconstruction are tracked as a function of time. For example, speckle or feature tracking (e.g., correlation or minimum sum of absolute differences) of the features is performed as a function of time. As another example, velocity estimates are used to indicate displacement. Other tracking techniques may be used. The amount of displacement, direction of displacement, and/or rotation of the features or two-dimensional image region defining a plane is determined. The plane position is adjusted based on the tracking. Images are generated in act 32 using the adjusted planes.
  • One example embodiment for determining an orientation of planes in a multi-planar reconstruction uses a pyramid of ultrasound data sets. More rapid determination may be provided. The ultrasound data is used to create two or more sets of data with different resolution (see FIG. 5 for a graphical example). For example, a one set of data has fine resolution, such as the scan resolution, and another set of data has a coarse resolution, such as the fine set decimated by ½ in each dimension. The sets represent the same object in the same volume. Three or more sets may be used.
  • The sets are in any format, such as Cartesian or polar coordinate. In one embodiment, the ultrasound data is acquired in an acoustic (e.g., polar) coordinate format, and the Cartesian or display space is populated in real time with only visible surfaces or selected planes. In another embodiment using a scan converter, processor, or graphics processing unit, real-time conversion from the acoustic space to the Cartesian or display space is feasible. The ultrasound data is processed in the Cartesian space (e.g., 3D grid) to orient the multi-planar reconstruction.
  • Given a volumetric pyramid, a hierarchical coarse-to-fine strategy is implemented. The data is processed on the highest level, such as scanning for feature points and/or planes. The subsequent levels are processed for further details. The pyramid processing may improve the processing speed. Alternatively, a single set of ultrasound data is used without the data pyramid.
  • The locations of multiple planes are detected. The 3D volumetric data is stored in the spherical coordinates. The relationship between the Cartesian coordinates (x, y, s) and the spherical coordinates (r, θ, Φ) is as follows:

  • x=r sin Φ sin θ; y=r sin Φ cos θ; z=r cos Φ
  • One parameterization of a plane in the Cartesian coordinates is:

  • x sin Φ0 sin θ0+y sin Φ0 cos θ0+z cos Φ0=r0
  • where (r0, θ0, Φ0) is the spherical coordinates of the perpendicular foot of the origin to the plane. Therefore, the plane in the spherical coordinates is parameterized as:

  • r sin Φ sin Φ0 cos (θ−θ0)+r cos Φ cos Φ0=r0
  • In a first stage, an estimate of the multiple planes is performed. Using the volume pyramid in the Cartesian space, the data is processed in a hierarchical manner to detect reliable feature points, such as an apex, valve annulus points, and/or chambers (e.g., the chamber of the left ventricle). The features provide a rough estimate of the positions of the multiple planes. Using the features, the whole heart or structure may be aligned to a canonical model.
  • Object features are identified. A processor identifies a plurality of features of the object from the first coarse set of ultrasound data. Larger features may be used for determining the locations in the coarse set. The features indicate the likely location of smaller features. The smaller features are then identified in the fine set of the volume pyramid based on the larger features. Alternatively, the smaller features are identified regardless of the larger features. The location of the larger features may be refined using the fine set with or without identifying other features in the fine set.
  • The processor determines locations of planes for multi-planar reconstruction as a function of the features of the object. As discussed above for act 30, the position of planes of the multi-planar reconstruction is located as a function of the object features. Images may then be generated.
  • One embodiment is for multi-plane detection and tracking in real-time 3D echocardiography. In this embodiment, the location for the multiple planes is initially determined without required user adjustment of an imaging or scanning plane. Once the positions are determined, the planes are tracked for subsequent imaging using a different approach.
  • In a second stage, detection for refinement may be programmed. In one embodiment, a learning approach to detection is used. An annotated database is obtained. Training positive and negative examples are extracted from the database.
  • A refined estimate of the positions of the multiple planes is determined from the set of data with a finer resolution. In the Cartesian space, the estimates of each plane positioning are refined. Any now known or later developed refinement may be used. In two possible embodiments, a designed object detector that discriminates the subtle difference in the appearance is trained, or a regressor is learned to infer the estimate. Alternatively, projection pursuit is used.
  • Points like the apex and valve annulus present distinct structures that may be well described by specified features. In addition, the heart chambers have semantic constraints that could be learned by higher-level classifiers. A hierarchical representation, modeling, and detection of the heart structures may be used.
  • In one approach, plane-specific detectors are learned. One detector for one particular plane may be learned. For example, an A4C detector and an A2C detector are learned separately. In another approach, a joint detector for multiple planes that are treated as an entity is learned.
  • The detector is a binary classifier F(I):
      • if F(I)≧0, then I is positive;
        • else, I is negative.
          The function F( I) also gives a confidence of being positive. The higher F(I) is, the more confident the input I is positive. The confidence may be interpreted as a posterior probability:
  • p ( 1 ) = exp ( F ( I ) ) exp ( F ( I ) ) + exp ( - F ( I ) )
  • The binary classifier known as the AdaBoosting algorithm, probabilistic boosting tree (PDT) or other classifiers may be used. With the ultrasound data, the boosting algorithm is used to select Haar features, which are amenable to rapid evaluation. The Haar features are combined to form a committee classifier. An image-based multi-class classifier to further differentiate multiple planes may be used in addition.
  • Given an ultrasound volume V and a configuration a, multiple planes denoted by I=V(a) are extracted. The optimal parameter a that maximizes the detector classifier p(I) is searched for where:
  • α = arg max α p ( V ( α ) )
  • When separate plane-specific classifiers are learned, the overall detector classifier p(I) is simply p(I)=p1(I)*p2(I)* . . . *pn(I), that is, a product of n quantities. Since the multiple planes of interest are geometrically constrained, some degrees of freedom are provided when performing the exhaustive search.
  • In projection pursuit, the posterior probability p(V(a)) is replaced by a pursuit cost function. For example, it can be based on projected histogram or other measures.
  • Regression may be used for refinement with the fine set of ultrasound data. One practical consideration is the searching speed especially when confronting the three-dimensional parameter space. To accelerate the speed, a function:

  • δα=g(I).
  • is learned, where δα is the difference between the current parameter α and the ground truth {acute over (α)}, as represented by:

  • δα=α−{umlaut over (α)}
  • In one embodiment, the locations of the planes are refined. The processor uses any function to refine the locations, such as a classifier, regressor, or combinations thereof. For example, from the annotated database, example pairs {(In, δαn)}n=1 N. An image-based regression task is used to find the target function g(I) that minimizes the cost function:
  • arg min g L ( g ) = n = 1 N L ( g ( I n ) , δα n ) + λ R ( g )
  • where L(g(ln), δαn) is the loss function and R(g) is a regularization term.
  • Based on the initial parameter a, the multiple slices I=V(α) are formed, and then the parameter δαn is inferred by applying the regression function. An updated estimate α10+δαn is obtained. By iterating the above process until convergence, the parameter a may be inferred reasonably close to the ground truth. The discriminative classifier is used to further tweak the estimate.
  • After detecting the planes from the ultrasound data, the planes are tracked. In real-time acquisition, after locking the multiple planes, the planes are tracked. Any tracking may be used. In an alternative embodiment, the detector is applied for every scan or ultrasound data representing the volume region.
  • Another approach is to use multiple hypotheses tracking. In multiple hypotheses tracking, hypotheses of a, are generated and verified by the filtering probability p(α1|V1:t). Tracking is performed in real-time. Due to sampling smoothness in the temporal dimension, the parameter cannot undergo a sudden change. Assume that a obeys a Markov model:

  • αtt−1+ut
  • where ut is a noise component. If the classifier p(I) is used as a likelihood measurement q(Vtt)=p(Vtt)), the posterior filtering probability p(αt|V1:t) is calculated. The optimal a, is estimated from the posterior filtering probability.
  • The Kalman filter may provide an exact solution. Often, the above time series system is not linear and the distributions are non-Gaussian. The sequential Monte Carlo (SMC) algorithm may be used to derive an approximate solution, allowing real-time tracking.
  • While the invention has been described above by reference to various embodiments, it should be understood that many changes and modifications can be made without departing from the scope of the invention. It is therefore intended that the foregoing detailed description be regarded as illustrative rather than limiting, and that it be understood that it is the following claims, including all equivalents, that are intended to define the spirit and scope of this invention.

Claims (23)

1. A method for multi-planar reconstruction for ultrasound volume data, the method comprising:
positioning an ultrasound transducer adjacent, on or within a patient;
scanning a volume region with the ultrasound transducer;
determining, with a processor and from data responsive to the scanning, a first orientation of an object within the volume region while scanning;
orienting a multi-planar reconstruction as a function of the first orientation of the object and independently of a second orientation of the ultrasound transducer relative to the object; and
generating multi-planar reconstruction images of the object from the data while scanning, the images being a function of the orientation of the multi-planar reconstruction.
2. The method of claim 1 wherein positioning comprises the user positioning a wobbler or multi-dimensional array, wherein scanning comprises scanning the volume region with the wobbler or multi-dimensional array, and wherein generating the multi-planar reconstruction comprises forming at least two two-dimensional images associated with different planes in the volume region.
3. The method of claim 1 wherein determining, generating and orienting are performed in real-time with the scanning.
4. The method of claim 1 wherein orienting the multi-planar reconstruction comprises positioning at least one plane of the reconstruction relative to the object such that the at least one plane corresponds to a standard view.
5. The method of claim 4 wherein the object comprises a heart, and wherein positioning comprises positioning the at least one plane for an apical two chamber, apical four chamber, parasternal long axis, or parasternal short axis view.
6. The method of claim 1 wherein orienting comprises orienting without fixing a view relative to the transducer and by fixing the view relative to the object.
7. The method of claim 1 further comprising:
repeating the determining, orienting, and generating a plurality of times while scanning in a same imaging session.
8. The method of claim 1 wherein orienting comprises orienting a plurality of planes to a respective plurality of views for the object.
9. The method of claim 1 further comprising:
tracking anatomical view planes of the multi-planar reconstruction as a function of time.
10. The method of claim 1 wherein generating comprises generating available standard views with the processor from the data for the object.
11. The method of claim 1 wherein determining comprises identifying a plurality of features of the object from the data in a first coarse set and in a second fine set of a volume pyramid.
12. The method of claim 1 wherein determining and orienting comprise:
identifying object features;
locating planes of the multi-planar reconstruction as a function of the object features; and
refining locations of the planes with a classifier, regressor, or combinations thereof.
13. The method of claim 1 further comprising:
tracking planes of the multi-planar reconstruction with multiple hypotheses tracking.
14. In a computer readable storage medium having stored therein data representing instructions executable by a programmed processor for multi-planar reconstruction for ultrasound volume data, the storage medium comprising instructions for:
controlling acquisition by scanning a volume region of a patient;
determining, from ultrasound data responsive to the acquisition, locations of features of an object within the volume region represented by the data, the determining being during control of the acquisition by scanning;
orienting a plurality of planes within the volume region as a function of the locations of the features, the orienting being independent of an orientation of an ultrasound transducer relative to the object, each of the plurality of planes being different from the other ones of the plurality of planes; and
generating images of the object from the data for each of the planes.
15. The instructions of claim 14 wherein orienting and generating are performed during the control of the acquisition by scanning.
16. The instructions of claim 14 wherein orienting comprises positioning the planes relative to the object such that the planes correspond to a predetermined views.
17. The instructions of claim 14 further comprising:
tracking the features as a function of time.
18. The instructions of claim 14 wherein determining comprises determining from the data at different resolutions.
19. The instructions of claim 14 further comprising:
refining locations of the planes with a classifier, regressor, or combinations thereof.
20. The instructions of claim 14 further comprising:
tracking the planes with multiple hypotheses tracking.
21. A method for multi-planar reconstruction for ultrasound volume data, the method comprising:
obtaining ultrasound data in a first coarse set and a second fine set, the ultrasound data representing an object in a volume;
identifying, with a processor, a plurality of features of the object from the first coarse set of ultrasound data;
determining, with the processor, locations of planes for multi-planar reconstruction as a function of the features of the object; and
refining, with the processor, the locations as a function of the second fine set.
22. The method of claim 21 wherein refining comprises refining with a classifier, regressor, or combinations thereof.
23. The method of claim 21 further comprising:
tracking the planes with multiple hypotheses tracking.
US11/527,286 2006-05-11 2006-09-25 Multi-planar reconstruction for ultrasound volume data Abandoned US20080009722A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/527,286 US20080009722A1 (en) 2006-05-11 2006-09-25 Multi-planar reconstruction for ultrasound volume data

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US74702406P 2006-05-11 2006-05-11
US11/527,286 US20080009722A1 (en) 2006-05-11 2006-09-25 Multi-planar reconstruction for ultrasound volume data

Publications (1)

Publication Number Publication Date
US20080009722A1 true US20080009722A1 (en) 2008-01-10

Family

ID=38919908

Family Applications (1)

Application Number Title Priority Date Filing Date
US11/527,286 Abandoned US20080009722A1 (en) 2006-05-11 2006-09-25 Multi-planar reconstruction for ultrasound volume data

Country Status (1)

Country Link
US (1) US20080009722A1 (en)

Cited By (35)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080021945A1 (en) * 2006-07-20 2008-01-24 James Hamilton Method of processing spatial-temporal data processing
US20080021319A1 (en) * 2006-07-20 2008-01-24 James Hamilton Method of modifying data acquisition parameters of an ultrasound device
US20080281203A1 (en) * 2007-03-27 2008-11-13 Siemens Corporation System and Method for Quasi-Real-Time Ventricular Measurements From M-Mode EchoCardiogram
US20090074280A1 (en) * 2007-09-18 2009-03-19 Siemens Corporate Research, Inc. Automated Detection of Planes From Three-Dimensional Echocardiographic Data
US20090304250A1 (en) * 2008-06-06 2009-12-10 Mcdermott Bruce A Animation for Conveying Spatial Relationships in Three-Dimensional Medical Imaging
US20090318803A1 (en) * 2008-06-19 2009-12-24 Yasuhiko Abe Ultrasonic diagnostic apparatus, ultrasonic image processing apparatus, and medical image processing apparatus
US20100041992A1 (en) * 2008-08-13 2010-02-18 Hiroyuki Ohuchi Ultrasonic diagnostic apparatus, ultrasonic image display apparatus, and medical image diagnostic apparatus
US20100086187A1 (en) * 2008-09-23 2010-04-08 James Hamilton System and method for flexible rate processing of ultrasound data
US20100125203A1 (en) * 2008-11-19 2010-05-20 Medison Co., Ltd. Finding A Standard View Corresponding To An Acquired Ultrasound Image
US20100138191A1 (en) * 2006-07-20 2010-06-03 James Hamilton Method and system for acquiring and transforming ultrasound data
US20100185085A1 (en) * 2009-01-19 2010-07-22 James Hamilton Dynamic ultrasound processing using object motion calculation
US20110087094A1 (en) * 2009-10-08 2011-04-14 Hiroyuki Ohuchi Ultrasonic diagnosis apparatus and ultrasonic image processing apparatus
US20110169864A1 (en) * 2008-09-26 2011-07-14 Koninklijke Philips Electronics N.V. Patient specific anatiomical sketches for medical reports
US20110263981A1 (en) * 2007-07-20 2011-10-27 James Hamilton Method for measuring image motion with synthetic speckle patterns
WO2012042449A3 (en) * 2010-09-30 2012-07-05 Koninklijke Philips Electronics N.V. Image and annotation display
WO2013056231A1 (en) * 2011-10-14 2013-04-18 Jointvue, Llc Real-time 3-d ultrasound reconstruction of knee and its complications for patient specific implants and 3-d joint injections
US20130324849A1 (en) * 2012-06-01 2013-12-05 Samsung Medison Co., Ltd. Method and apparatus for displaying ultrasonic image and information related to the ultrasonic image
US20130331697A1 (en) * 2012-06-11 2013-12-12 Samsung Medison Co., Ltd. Method and apparatus for displaying three-dimensional ultrasonic image and two-dimensional ultrasonic image
US20130338496A1 (en) * 2010-12-13 2013-12-19 The Trustees Of Columbia University In The City New York Medical imaging devices, methods, and systems
US20140187919A1 (en) * 2011-04-21 2014-07-03 Koninklijke Philips N.V. Mpr slice selection for visualization of catheter in three-dimensional ultrasound
US9025858B2 (en) 2011-01-25 2015-05-05 Samsung Electronics Co., Ltd. Method and apparatus for automatically generating optimal 2-dimensional medical image from 3-dimensional medical image
EP2893880A1 (en) * 2014-01-08 2015-07-15 Samsung Medison Co., Ltd. Ultrasound diagnostic apparatus and method of operating the same
US20150302638A1 (en) * 2012-11-20 2015-10-22 Koninklijke Philips N.V Automatic positioning of standard planes for real-time fetal heart evaluation
US20160027184A1 (en) * 2013-03-15 2016-01-28 Colibri Technologies Inc. Data display and processing algorithms for 3d imaging systems
WO2016156481A1 (en) * 2015-03-31 2016-10-06 Koninklijke Philips N.V. Ultrasound imaging apparatus
US9642572B2 (en) 2009-02-02 2017-05-09 Joint Vue, LLC Motion Tracking system with inertial-based sensing units
EP3363365A1 (en) * 2011-12-12 2018-08-22 Koninklijke Philips N.V. Automatic imaging plane selection for echocardiography
US20190145940A1 (en) * 2017-11-14 2019-05-16 Ge Sensing & Inspection Technologies Gmbh Classification of Ultrasonic Indications Using Pattern Recognition
US10321892B2 (en) * 2010-09-27 2019-06-18 Siemens Medical Solutions Usa, Inc. Computerized characterization of cardiac motion in medical diagnostic ultrasound
US10394416B2 (en) * 2013-12-31 2019-08-27 Samsung Electronics Co., Ltd. User interface system and method for enabling mark-based interaction for images
US10512451B2 (en) 2010-08-02 2019-12-24 Jointvue, Llc Method and apparatus for three dimensional reconstruction of a joint using ultrasound
US10517568B2 (en) 2011-08-12 2019-12-31 Jointvue, Llc 3-D ultrasound imaging device and methods
US20210128108A1 (en) * 2019-11-05 2021-05-06 Siemens Medical Solutions Usa, Inc. Loosely coupled probe position and view in ultrasound imaging
US20220304651A1 (en) * 2021-03-25 2022-09-29 Canon Medical Systems Corporation Ultrasound diagnostic apparatus, medical image analytic apparatus, and non-transitory computer readable storage medium storing medical image analysis program
EP3791779B1 (en) * 2013-02-04 2022-12-28 Jointvue, LLC Method for 3d reconstruction of a joint using ultrasound

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6464642B1 (en) * 1999-08-20 2002-10-15 Kabushiki Kaisha Toshiba Ultrasonic diagnosis apparatus
US6589176B2 (en) * 2001-12-05 2003-07-08 Koninklijke Philips Electronics N.V. Ultrasonic image stabilization system and method
US6733458B1 (en) * 2001-09-25 2004-05-11 Acuson Corporation Diagnostic medical ultrasound systems and methods using image based freehand needle guidance
US6768820B1 (en) * 2000-06-06 2004-07-27 Agilent Technologies, Inc. Method and system for extracting data from surface array deposited features
US20040254439A1 (en) * 2003-06-11 2004-12-16 Siemens Medical Solutions Usa, Inc. System and method for adapting the behavior of a diagnostic medical ultrasound system based on anatomic features present in ultrasound images
US20050096538A1 (en) * 2003-10-29 2005-05-05 Siemens Medical Solutions Usa, Inc. Image plane stabilization for medical imaging
US20050100203A1 (en) * 2003-11-10 2005-05-12 Yasuko Fujisawa Image processor
US20050105788A1 (en) * 2003-11-19 2005-05-19 Matthew William Turek Methods and apparatus for processing image data to aid in detecting disease
US20060034513A1 (en) * 2004-07-23 2006-02-16 Siemens Medical Solutions Usa, Inc. View assistance in three-dimensional ultrasound imaging

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6464642B1 (en) * 1999-08-20 2002-10-15 Kabushiki Kaisha Toshiba Ultrasonic diagnosis apparatus
US6768820B1 (en) * 2000-06-06 2004-07-27 Agilent Technologies, Inc. Method and system for extracting data from surface array deposited features
US6733458B1 (en) * 2001-09-25 2004-05-11 Acuson Corporation Diagnostic medical ultrasound systems and methods using image based freehand needle guidance
US6589176B2 (en) * 2001-12-05 2003-07-08 Koninklijke Philips Electronics N.V. Ultrasonic image stabilization system and method
US20040254439A1 (en) * 2003-06-11 2004-12-16 Siemens Medical Solutions Usa, Inc. System and method for adapting the behavior of a diagnostic medical ultrasound system based on anatomic features present in ultrasound images
US7092749B2 (en) * 2003-06-11 2006-08-15 Siemens Medical Solutions Usa, Inc. System and method for adapting the behavior of a diagnostic medical ultrasound system based on anatomic features present in ultrasound images
US20050096538A1 (en) * 2003-10-29 2005-05-05 Siemens Medical Solutions Usa, Inc. Image plane stabilization for medical imaging
US20050100203A1 (en) * 2003-11-10 2005-05-12 Yasuko Fujisawa Image processor
US20050105788A1 (en) * 2003-11-19 2005-05-19 Matthew William Turek Methods and apparatus for processing image data to aid in detecting disease
US20060034513A1 (en) * 2004-07-23 2006-02-16 Siemens Medical Solutions Usa, Inc. View assistance in three-dimensional ultrasound imaging

Cited By (69)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080021319A1 (en) * 2006-07-20 2008-01-24 James Hamilton Method of modifying data acquisition parameters of an ultrasound device
US20080021945A1 (en) * 2006-07-20 2008-01-24 James Hamilton Method of processing spatial-temporal data processing
US20100138191A1 (en) * 2006-07-20 2010-06-03 James Hamilton Method and system for acquiring and transforming ultrasound data
US20080281203A1 (en) * 2007-03-27 2008-11-13 Siemens Corporation System and Method for Quasi-Real-Time Ventricular Measurements From M-Mode EchoCardiogram
US8396531B2 (en) * 2007-03-27 2013-03-12 Siemens Medical Solutions Usa, Inc. System and method for quasi-real-time ventricular measurements from M-mode echocardiogram
US9275471B2 (en) * 2007-07-20 2016-03-01 Ultrasound Medical Devices, Inc. Method for ultrasound motion tracking via synthetic speckle patterns
US20110263981A1 (en) * 2007-07-20 2011-10-27 James Hamilton Method for measuring image motion with synthetic speckle patterns
US20090074280A1 (en) * 2007-09-18 2009-03-19 Siemens Corporate Research, Inc. Automated Detection of Planes From Three-Dimensional Echocardiographic Data
US8073215B2 (en) 2007-09-18 2011-12-06 Siemens Medical Solutions Usa, Inc. Automated detection of planes from three-dimensional echocardiographic data
US20090304250A1 (en) * 2008-06-06 2009-12-10 Mcdermott Bruce A Animation for Conveying Spatial Relationships in Three-Dimensional Medical Imaging
US8494250B2 (en) 2008-06-06 2013-07-23 Siemens Medical Solutions Usa, Inc. Animation for conveying spatial relationships in three-dimensional medical imaging
US20090318803A1 (en) * 2008-06-19 2009-12-24 Yasuhiko Abe Ultrasonic diagnostic apparatus, ultrasonic image processing apparatus, and medical image processing apparatus
US9186125B2 (en) * 2008-06-19 2015-11-17 Kabushiki Kaisha Toshiba Ultrasonic diagnostic apparatus for generating three dimensional cardiac motion image by setting line segmented strain gauges
US20100041992A1 (en) * 2008-08-13 2010-02-18 Hiroyuki Ohuchi Ultrasonic diagnostic apparatus, ultrasonic image display apparatus, and medical image diagnostic apparatus
US10792009B2 (en) * 2008-08-13 2020-10-06 Canon Medical Systems Corporation Ultrasonic diagnostic apparatus, ultrasonic image display apparatus, and medical image diagnostic apparatus
US20100086187A1 (en) * 2008-09-23 2010-04-08 James Hamilton System and method for flexible rate processing of ultrasound data
US20110169864A1 (en) * 2008-09-26 2011-07-14 Koninklijke Philips Electronics N.V. Patient specific anatiomical sketches for medical reports
US10102347B2 (en) * 2008-09-26 2018-10-16 Koninklijke Philips N.V. Patient specific anatiomical sketches for medical reports
EP2189807A3 (en) * 2008-11-19 2010-12-08 Medison Co., Ltd. Finding a standard view corresponding to an acquired ultrasound image
EP2189807A2 (en) * 2008-11-19 2010-05-26 Medison Co., Ltd. Finding a standard view corresponding to an acquired ultrasound image
KR101051567B1 (en) * 2008-11-19 2011-07-22 삼성메디슨 주식회사 Ultrasound systems and methods that provide standard cross section information
US20100125203A1 (en) * 2008-11-19 2010-05-20 Medison Co., Ltd. Finding A Standard View Corresponding To An Acquired Ultrasound Image
US20100185085A1 (en) * 2009-01-19 2010-07-22 James Hamilton Dynamic ultrasound processing using object motion calculation
US11004561B2 (en) 2009-02-02 2021-05-11 Jointvue Llc Motion tracking system with inertial-based sensing units
US11342071B2 (en) 2009-02-02 2022-05-24 Jointvue, Llc Noninvasive diagnostic system
US9642572B2 (en) 2009-02-02 2017-05-09 Joint Vue, LLC Motion Tracking system with inertial-based sensing units
JP2011078625A (en) * 2009-10-08 2011-04-21 Toshiba Corp Ultrasonic diagnosis apparatus, ultrasonic image processing apparatus, and ultrasonic image processing program
US20110087094A1 (en) * 2009-10-08 2011-04-14 Hiroyuki Ohuchi Ultrasonic diagnosis apparatus and ultrasonic image processing apparatus
US10512451B2 (en) 2010-08-02 2019-12-24 Jointvue, Llc Method and apparatus for three dimensional reconstruction of a joint using ultrasound
US10321892B2 (en) * 2010-09-27 2019-06-18 Siemens Medical Solutions Usa, Inc. Computerized characterization of cardiac motion in medical diagnostic ultrasound
RU2598329C2 (en) * 2010-09-30 2016-09-20 Конинклейке Филипс Электроникс Н.В. Displaying images and annotations
WO2012042449A3 (en) * 2010-09-30 2012-07-05 Koninklijke Philips Electronics N.V. Image and annotation display
US9514575B2 (en) 2010-09-30 2016-12-06 Koninklijke Philips N.V. Image and annotation display
KR20140058399A (en) 2010-12-13 2014-05-14 더 트러스티이스 오브 콜롬비아 유니버시티 인 더 시티 오브 뉴욕 Medical imaging devices, methods, and systems
US20130338496A1 (en) * 2010-12-13 2013-12-19 The Trustees Of Columbia University In The City New York Medical imaging devices, methods, and systems
US9486142B2 (en) * 2010-12-13 2016-11-08 The Trustees Of Columbia University In The City Of New York Medical imaging devices, methods, and systems
KR101935064B1 (en) * 2010-12-13 2019-03-18 더 트러스티이스 오브 콜롬비아 유니버시티 인 더 시티 오브 뉴욕 Medical imaging devices, methods, and systems
US9025858B2 (en) 2011-01-25 2015-05-05 Samsung Electronics Co., Ltd. Method and apparatus for automatically generating optimal 2-dimensional medical image from 3-dimensional medical image
US10376179B2 (en) * 2011-04-21 2019-08-13 Koninklijke Philips N.V. MPR slice selection for visualization of catheter in three-dimensional ultrasound
US20140187919A1 (en) * 2011-04-21 2014-07-03 Koninklijke Philips N.V. Mpr slice selection for visualization of catheter in three-dimensional ultrasound
US10517568B2 (en) 2011-08-12 2019-12-31 Jointvue, Llc 3-D ultrasound imaging device and methods
WO2013056231A1 (en) * 2011-10-14 2013-04-18 Jointvue, Llc Real-time 3-d ultrasound reconstruction of knee and its complications for patient specific implants and 3-d joint injections
US20140221825A1 (en) * 2011-10-14 2014-08-07 Jointvue, Llc Real-Time 3-D Ultrasound Reconstruction of Knee and Its Implications For Patient Specific Implants and 3-D Joint Injections
US11123040B2 (en) 2011-10-14 2021-09-21 Jointvue, Llc Real-time 3-D ultrasound reconstruction of knee and its implications for patient specific implants and 3-D joint injections
US20210378631A1 (en) * 2011-10-14 2021-12-09 Jointvue, Llc Real-Time 3-D Ultrasound Reconstruction of Knee and Its Implications For Patient Specific Implants and 3-D Joint Injections
US11529119B2 (en) * 2011-10-14 2022-12-20 Jointvue, Llc Real-time 3-D ultrasound reconstruction of knee and its implications for patient specific implants and 3-D joint injections
US11819359B2 (en) * 2011-10-14 2023-11-21 Jointvue, Llc Real-time 3-D ultrasound reconstruction of knee and its implications for patient specific implants and 3-D joint injections
EP3363365A1 (en) * 2011-12-12 2018-08-22 Koninklijke Philips N.V. Automatic imaging plane selection for echocardiography
US20130324849A1 (en) * 2012-06-01 2013-12-05 Samsung Medison Co., Ltd. Method and apparatus for displaying ultrasonic image and information related to the ultrasonic image
US20130331697A1 (en) * 2012-06-11 2013-12-12 Samsung Medison Co., Ltd. Method and apparatus for displaying three-dimensional ultrasonic image and two-dimensional ultrasonic image
US20150302638A1 (en) * 2012-11-20 2015-10-22 Koninklijke Philips N.V Automatic positioning of standard planes for real-time fetal heart evaluation
US9734626B2 (en) * 2012-11-20 2017-08-15 Koninklijke Philips N.V. Automatic positioning of standard planes for real-time fetal heart evaluation
US10410409B2 (en) 2012-11-20 2019-09-10 Koninklijke Philips N.V. Automatic positioning of standard planes for real-time fetal heart evaluation
RU2654611C2 (en) * 2012-11-20 2018-05-21 Конинклейке Филипс Н.В. Automatic positioning of standard planes for real-time fetal heart evaluation
EP3791779B1 (en) * 2013-02-04 2022-12-28 Jointvue, LLC Method for 3d reconstruction of a joint using ultrasound
US20160027184A1 (en) * 2013-03-15 2016-01-28 Colibri Technologies Inc. Data display and processing algorithms for 3d imaging systems
AU2014231354B2 (en) * 2013-03-15 2019-08-29 Conavi Medical Inc. Data display and processing algorithms for 3D imaging systems
US9786056B2 (en) * 2013-03-15 2017-10-10 Sunnybrook Research Institute Data display and processing algorithms for 3D imaging systems
US10394416B2 (en) * 2013-12-31 2019-08-27 Samsung Electronics Co., Ltd. User interface system and method for enabling mark-based interaction for images
EP2893880A1 (en) * 2014-01-08 2015-07-15 Samsung Medison Co., Ltd. Ultrasound diagnostic apparatus and method of operating the same
WO2016156481A1 (en) * 2015-03-31 2016-10-06 Koninklijke Philips N.V. Ultrasound imaging apparatus
JP2021119990A (en) * 2015-03-31 2021-08-19 コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. Ultrasound imaging apparatus
US11006927B2 (en) 2015-03-31 2021-05-18 Koninklijke Philips N.V. Ultrasound imaging apparatus
CN107405134A (en) * 2015-03-31 2017-11-28 皇家飞利浦有限公司 Supersonic imaging device
JP7216140B2 (en) 2015-03-31 2023-01-31 コーニンクレッカ フィリップス エヌ ヴェ Ultrasound imaging device
US10648951B2 (en) * 2017-11-14 2020-05-12 Ge Sensing & Inspection Technologies Gmbh Classification of ultrasonic indications using pattern recognition
US20190145940A1 (en) * 2017-11-14 2019-05-16 Ge Sensing & Inspection Technologies Gmbh Classification of Ultrasonic Indications Using Pattern Recognition
US20210128108A1 (en) * 2019-11-05 2021-05-06 Siemens Medical Solutions Usa, Inc. Loosely coupled probe position and view in ultrasound imaging
US20220304651A1 (en) * 2021-03-25 2022-09-29 Canon Medical Systems Corporation Ultrasound diagnostic apparatus, medical image analytic apparatus, and non-transitory computer readable storage medium storing medical image analysis program

Similar Documents

Publication Publication Date Title
US20080009722A1 (en) Multi-planar reconstruction for ultrasound volume data
US8073215B2 (en) Automated detection of planes from three-dimensional echocardiographic data
US7648460B2 (en) Medical diagnostic imaging optimization based on anatomy recognition
US8092388B2 (en) Automated view classification with echocardiographic data for gate localization or other purposes
US9033887B2 (en) Mitral valve detection for transthoracic echocardiography
KR101907550B1 (en) Needle enhancement in diagnostic ultrasound imaging
US8556814B2 (en) Automated fetal measurement from three-dimensional ultrasound data
US10271817B2 (en) Valve regurgitant detection for echocardiography
US10321892B2 (en) Computerized characterization of cardiac motion in medical diagnostic ultrasound
US10123781B2 (en) Automated segmentation of tri-plane images for real time ultrasonic imaging
US20220079552A1 (en) Cardiac flow detection based on morphological modeling in medical diagnostic ultrasound imaging
US9179890B2 (en) Model-based positioning for intracardiac echocardiography volume stitching
CN110719755B (en) Ultrasonic imaging method
JP2020527080A (en) Fetal ultrasound image processing
US11712224B2 (en) Method and systems for context awareness enabled ultrasound scanning
US20060239527A1 (en) Three-dimensional cardiac border delineation in medical imaging
CN111683600A (en) Apparatus and method for obtaining anatomical measurements from ultrasound images
US9033883B2 (en) Flow quantification in ultrasound using conditional random fields with global consistency
JP2022111140A (en) Ultrasound diagnosis apparatus
US11717268B2 (en) Ultrasound imaging system and method for compounding 3D images via stitching based on point distances
JP2023552330A (en) Predicting the likelihood that an individual will have one or more diseases
Zhou et al. Artificial intelligence in quantitative ultrasound imaging: A review
US20220370046A1 (en) Robust view classification and measurement in ultrasound imaging
CN116033874A (en) System and method for measuring cardiac stiffness

Legal Events

Date Code Title Description
AS Assignment

Owner name: SIEMENS MEDICAL SOLUTIONS USA, INC., PENNSYLVANIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SIMOPOULOS, CONSTANTINE;THOMAS III, LEWIS J.;ZHOU, SHAOHUA KEVIN;AND OTHERS;REEL/FRAME:018913/0657;SIGNING DATES FROM 20061130 TO 20061215

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION