US5575286A - Method and apparatus for generating large compound ultrasound image - Google Patents

Method and apparatus for generating large compound ultrasound image Download PDF

Info

Publication number
US5575286A
US5575286A US08/414,978 US41497895A US5575286A US 5575286 A US5575286 A US 5575286A US 41497895 A US41497895 A US 41497895A US 5575286 A US5575286 A US 5575286A
Authority
US
United States
Prior art keywords
image
local motion
motion vectors
estimation
estimating
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
US08/414,978
Inventor
Lee Weng
Arun P. Tirumalai
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Siemens Medical Solutions USA Inc
Original Assignee
Siemens Medical Systems Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Siemens Medical Systems Inc filed Critical Siemens Medical Systems Inc
Priority to US08/414,978 priority Critical patent/US5575286A/en
Assigned to SIEMENS MEDICAL SYSTEMS, INC. reassignment SIEMENS MEDICAL SYSTEMS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: TIRUMALAI, CHUN Q., WENG, LEE
Priority to DE19611990A priority patent/DE19611990C2/en
Priority to JP07924396A priority patent/JP3345257B2/en
Priority to US08/728,280 priority patent/US5899861A/en
Application granted granted Critical
Priority to US08/747,429 priority patent/US5782766A/en
Publication of US5575286A publication Critical patent/US5575286A/en
Assigned to SIEMENS MEDICAL SOLUTIONS USA, INC. reassignment SIEMENS MEDICAL SOLUTIONS USA, INC. CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: SIEMENS MEDICAL SYSTEMS, INC.
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/52Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
    • G01S7/52017Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00 particularly adapted to short-range imaging
    • G01S7/52053Display arrangements
    • G01S7/52057Cathode ray tube displays
    • G01S7/52074Composite displays, e.g. split-screen displays; Combination of multiple images or of images and alphanumeric tabular information
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/08Detecting organic movements or changes, e.g. tumours, cysts, swellings
    • A61B8/0866Detecting organic movements or changes, e.g. tumours, cysts, swellings involving foetal diagnosis; pre-natal or peri-natal diagnosis of the baby
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/46Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient
    • A61B8/461Displaying means of special interest
    • A61B8/463Displaying means of special interest characterised by displaying multiple images or images and diagnostic data on one display
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/48Diagnostic techniques
    • A61B8/483Diagnostic techniques involving the acquisition of a 3D volume of data
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S15/00Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
    • G01S15/88Sonar systems specially adapted for specific applications
    • G01S15/89Sonar systems specially adapted for specific applications for mapping or imaging
    • G01S15/8906Short-range imaging systems; Acoustic microscope systems using pulse-echo techniques
    • G01S15/8979Combined Doppler and pulse-echo imaging systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/223Analysis of motion using block-matching
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/223Analysis of motion using block-matching
    • G06T7/238Analysis of motion using block-matching using non-full search, e.g. three-step search
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/08Detecting organic movements or changes, e.g. tumours, cysts, swellings
    • A61B8/0833Detecting organic movements or changes, e.g. tumours, cysts, swellings involving detecting or locating foreign bodies or organic structures
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S15/00Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
    • G01S15/02Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems using reflection of acoustic waves
    • G01S15/50Systems of measurement, based on relative movement of the target
    • G01S15/52Discriminating between fixed and moving objects or between objects moving at different speeds
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S15/00Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
    • G01S15/88Sonar systems specially adapted for specific applications
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S15/00Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
    • G01S15/88Sonar systems specially adapted for specific applications
    • G01S15/89Sonar systems specially adapted for specific applications for mapping or imaging
    • G01S15/8906Short-range imaging systems; Acoustic microscope systems using pulse-echo techniques
    • G01S15/8993Three dimensional imaging systems
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/52Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
    • G01S7/52017Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00 particularly adapted to short-range imaging
    • G01S7/52053Display arrangements
    • G01S7/52057Cathode ray tube displays
    • G01S7/5206Two-dimensional coordinated display of distance and direction; B-scan display
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/52Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
    • G01S7/52017Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00 particularly adapted to short-range imaging
    • G01S7/52053Display arrangements
    • G01S7/52057Cathode ray tube displays
    • G01S7/52068Stereoscopic displays; Three-dimensional displays; Pseudo 3D displays
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/52Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
    • G01S7/52017Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00 particularly adapted to short-range imaging
    • G01S7/52053Display arrangements
    • G01S7/52057Cathode ray tube displays
    • G01S7/52073Production of cursor lines, markers or indicia by electronic means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10068Endoscopic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person

Definitions

  • This invention relates to ultrasonic diagnostic imaging, and more particularly, to an ultrasonic imaging system that produces a large compound image by correlating movement between consecutive image frames.
  • Ultrasonic imaging techniques are commonly used to produce two-dimensional diagnostic images of internal features of an object, such as a human anatomy.
  • a diagnostic ultrasonic imaging system for medical use forms images of internal tissues of a human body by electrically exciting an acoustic transducer element or an array of acoustic transducer elements to generate short ultrasonic pulses that travel into the body.
  • the ultrasonic pulses produce echoes as they reflect off of body tissues that appear as discontinuities or impedance changes to the propagating ultrasonic pulses. These echoes return to the transducer, and are converted back into electrical signals that are amplified and decoded to produce a cross-sectional image of the tissues.
  • These ultrasonic imaging systems are of significant importance to the medical field by providing physicians with real-time, high resolution images of the internal features of a human anatomy without resort to more invasive exploratory techniques, such as surgery.
  • the acoustic transducer which radiates the ultrasonic pulses typically comprises a piezoelectric element or matrix of piezoelectric elements.
  • a piezoelectric element deforms upon application of an electrical signal to produce the ultrasonic pulses.
  • the received echoes cause the piezoelectric element to deform and generate the corresponding electrical signal.
  • the acoustic transducer is often packaged within a handheld device that allows the physician substantial freedom to easily manipulate the transducer over a desired area of interest.
  • the transducer can then be electrically connected via a cable to a central control device that generates and processes the electrical signals.
  • the control device transmits the image information to a real-time viewing device, such as a video display terminal.
  • the image information may also be stored so that other physicians may view the diagnostic images at a later date.
  • the individual images produced by such ultrasonic imaging systems comprise discrete frames, with each such frame having a field of view limited by the relatively narrow region traversed by the ultrasonic pulses.
  • each previous image is replaced on the viewing device by a new image defined by the limited field of view of the transducer.
  • a skilled physician can usually interpret the discrete frames in order to obtain a clear mental picture of the entire region traversed by the transducer, the discrete frames cannot be easily tiled together to produce a single, contiguous image. This can represent a significant drawback of conventional ultrasonic imaging systems, since it is not always possible for the physician to fully appreciate the overall condition of the body by consideration of the discrete frames alone. In some extreme cases, important information concerning the condition of the body tissues can be overlooked with serious potential consequences for the patient.
  • B-scanners utilize a transducer mounted on an arm assembly that constrains the transducer to move along a single plane or axis. Either the arm assembly or the transducer element itself can be provided with sensing devices that track the precise position of the transducer. This positional information could then be utilized to register each of the discrete image frames together into a single composite image.
  • An example of a compound B-scanner utilizing angular sensing devices on an arm assembly is disclosed in U.S. Pat. No.
  • This application provides a method and apparatus for generating a large compound ultrasonic image.
  • the method and apparatus utilizes image registration to eliminate the need for positional sensing devices altogether.
  • image registration method and apparatus consecutive moving image frames are correlated in order to derive the transducer motion information for compound image generation.
  • the method and apparatus recognizes that ultrasonic images from real-time scanning operations are highly correlated from one frame to another. Based on this recognition, the individual images are divided into several smaller sub-image regions, and a very fast and robust image motion detection algorithm is utilized to measure the sub-image motion.
  • This algorithm uses a fast adaptive coarse/fine minimum-sum-absolute-difference (MSAD) search strategy to compute an initial estimation of the local motion vectors. The initial estimation is combined with two measurement parameters by use of a fuzzy logic technique to derive a final estimation of the local motion vectors. The final local motion vectors are applied to a least-squares (L-S) process in order to estimate global image motion.
  • L-S least-squares
  • a fast display technique generates compound images based on the derived global image motion.
  • FIG. 1 is a partial perspective view of an ultrasonic imaging system adapted for real-time display of an exemplary fetal anatomy within a clinical environment;
  • FIG. 2 is a block diagram illustrating an image registration method in accordance with this invention
  • FIG. 3 is a block diagram illustrating a minimum-sum-absolute-difference search of the image registration method
  • FIG. 4 is a block diagram illustrating an alternative embodiment of the minimum-sum-absolute-difference search
  • FIG. 5 is a three-dimensional graph illustrating derivation of a quality factor for the minimum-sum-absolute-difference search
  • FIG. 6 is a block diagram illustrating derivation of a local vector deviation factor for the minimum-sum-absolute-difference search.
  • FIGS. 7A through 7D are graphs illustrating fuzzy logic membership functions for the quality factor and the deviation factor to derive a control value and a single numerical output.
  • This invention satisfies the critical need for a method and apparatus to combine each of the discrete image frames produced by an ultrasonic imaging system into a single composite image.
  • consecutive image frames are correlated in order to derive the transducer motion information for compound imaging.
  • the motion information is derived without encumbering the handheld transducer with position sensing devices that would otherwise increase the cost, weight and complexity of such ultrasonic imaging systems.
  • FIG. 1 a prior art ultrasonic imaging system adapted for real-time display within a clinical environment is illustrated.
  • a patient 10 carrying a fetus 12 is lying in a supine position to permit a physician or other technician 14 to perform an ultrasonic imaging operation.
  • the imaging system comprises a scanhead 16 coupled to an ultrasound scanner 18.
  • the physician 14 draws the scanhead 16 along the patient's skin surface in a direction which is parallel with the ultrasonic scan plane.
  • An acoustically conductive lubricating agent may be applied to the skin prior to bringing the scanhead 16 into contact with the skin so as to improve acoustic coupling between the scanhead and the patient.
  • the scanhead 16 includes an ultrasonic transducer disposed at a surface thereof, comprising a piezoelectric element or matrix of individual piezoelectric elements.
  • the ultrasound scanner 18 provides electrical signals to the scanhead 16 that cause it to generate ultrasonic pulses.
  • the ultrasonic pulses 24 propagate in the scan plane through the skin of the patient 10 and echo off of the anatomical features of the fetus 12.
  • the echoes return through the skin to the scanhead 16, which converts the echoes back into electrical signals received by the ultrasound scanner 18.
  • the received electrical signals are then transmitted from the ultrasound scanner 18 to an image processor 20.
  • the image processor 20 decodes the electrical signals into a two-dimensional, cross-sectional image that is displayed on the video display terminal 22.
  • the image information may also be electronically stored in a memory medium, including a permanent storage medium such as a disk or tape drive, a temporary storage medium such as a solid state memory, or can be printed to a hard copy image such as a photograph.
  • the ultrasonic image displayed on the video display terminal 22 (or stored by use of the various other storage media described above) comprises only the information representative of the relatively narrow region traversed by the ultrasonic pulses 24. As illustrated in FIG. 1, only a portion of the anatomical features of the fetus 12 are displayed on the video display terminal 22. The region outside the sector of the ultrasonic pulses 24 (illustrated in phantom in FIG. 1) actually appears blank on the display device 22. As noted above, it would be very desirable to provide a large compound image, such as a single image that includes the entire image region 26.
  • FIG. 2 a block diagram of an image registration method in accordance with this invention is illustrated.
  • Two consecutive image frames 31, 32 are denoted as frame n-1 (the previous frame) and frame n (the current frame), respectively.
  • the previous image frame n-1 is treated as a reference image.
  • the current image frame n acquired by moving the transducer 16 of FIG. 1 along the skin surface is compared with the previous image frame n-1 to estimate motion of the transducer.
  • frame n is divided into a plurality of sub-frame regions 34, and the local motion of each of the sub-frame regions is estimated.
  • the i'th block of frame n is defined as n(i).
  • the particular size and number of the blocks 34 are selected by consideration of various factors.
  • the block size should approximate the size of the particular image features. For example, image features such as tissue boundaries and blood vessels would require a relatively small block size. Nevertheless, such a small block size would not be suitable for estimation of motion because the small block size decorrelates very rapidly with relatively large magnitudes of motion. Moreover, the small block size would not be stable for in-vivo images where small scale tissue motion exists. At the same time, if the block size is too large, there will be too few blocks per image frame for motion estimation and the motion estimation will be unstable. Further, a large block size may introduce an image rotation error into the local motion estimation where the block is incorrectly assumed to have a translation component but no rotation component.
  • a block size of approximately 48 ⁇ 48 pixels is deemed acceptable for motion estimation in view of these factors in relation to an image frame size of about 200 ⁇ 400 pixels.
  • a minimum block size of 32 ⁇ 32 pixels could also be utilized in relation to a smaller image size.
  • each block n(i) of image n is estimated by moving the block n(i) (such as exemplary block 34) around on the n-1 image frame to find the best match.
  • the search region size 33 on the n-1 image frame is less than the total area of the frame 31, and should relate to the expected image motion speed and image frame rate. For example, a large search region size 33 should be used when the transducer speed is high and also when the image frame rate is low.
  • a search region size of 32 pixels in each of the four searching directions is adequate for use with an input image frame rate selected at 7.5 frame/second (computing one out of every four video frames), resulting in a search region size of 64 ⁇ 64 pixels.
  • the "best match” is determined by use of a minimum-sum-absolute-difference technique (MSAD) to match the n(i) block 34 with the search region 33 of frame n-1.
  • the sum-absolute-difference (SAD) is the sum of absolute differences between corresponding pixel values between each block 34 and the search region 33.
  • the "best match" between the block 34 and the search region 33 occurs where the SAD value is at a minimum.
  • FIG. 5 illustrates a SAD map having values that extend in the X and Y directions.
  • the generally uniform SAD map contains a valley where the SAD values dip below the mean, representing the location of the MSAD.
  • the MSAD technique is used to derive a first estimation of local motion vector v(i) having a direction and magnitude indicating how the block n(i) translated from the n-1 image frame to the n image frame.
  • the direction of the local motion vector v(i) is denoted by the arrow of FIG. 3.
  • calculating the MSAD could be a very slow computing process if the number of blocks is large, the block size is large, and the search region is large. For example, a frame n having sixteen blocks 34 with a 48 ⁇ 48 pixel block size and a search region of 64 ⁇ 64 pixels would require about 300,000,000 separate additions/subtractions and about 65,000 comparisons to complete a single MSAD computation. Even with the relatively high speed of conventional processors, this would still encompass too much computation to accommodate real-time application. Accordingly, various search techniques are utilized in order to reduce the magnitude of the MSAD computation.
  • MPEG Moving Picture Expert Group
  • MPEG encoders are commercially available that can be used to perform a rough estimate of MSAD location based on historical movement of the image.
  • a "best match" is sought between an image characteristic within a block 34 and a search region by searching within a local neighborhood comprising a finite dimensional range, such as within ten pixels in the X and Y directions relative to the image characteristic.
  • the location of the MSAD is identified by application of the MPEG encoder, and a local motion vector assigned to the block 34.
  • a drawback of the use of an MPEG encoder is that it results in a high incidence of inaccurate local motion vectors, which must be filtered out.
  • the accuracy of the local motion vector determination can be further improved by consideration of certain unique constraints of compound ultrasonic imaging.
  • Motion in the Y direction (vertical) from one frame to another is almost always smaller than the motion in the X direction (horizontal), because the ultrasonic transducer 16 generally moves only along the patient body contour.
  • This transducer motion can be characterized as being more of a horizontal extension than an up/down folding. Further, the transducer motion is most likely to be in one direction and continuous; sudden reversals in motion direction are not very likely. It has been found from a large number of experiments that MSAD maps are often fairly smooth and continuous even for blocks which have a lot of noise and lack of image features.
  • a fast adaptive coarse/fine MSAD search strategy can be devised that significantly reduces the total amount of computation.
  • the search strategy requires that the search range in the X direction be larger than that in the Y direction, and the search range and direction should be adaptive to the overall motion history. For example, if the history shows the image motion is in +X direction, it is very likely the subsequent frame will also move in the same direction.
  • the MSAD map is smooth and continuous, a coarse search can be performed first to reduce the search region 33 to a smaller area, followed by a fine search within the smaller area. Because motion in the Y direction is usually very small (and usually zero), the two-dimensional MSAD search can be reduced to two one-dimensional searches in the X and Y directions, respectively.
  • the first search should be conducted in the X direction to quickly narrow down the search region, followed by subsequent alternating one-dimensional searches in both the X and Y directions to quickly find the MSAD location. Points which have been searched during the coarse search or have been searched in the other direction could be skipped during the fine search.
  • the location of the MSAD can be identified after one coarse and one medium search in the X direction, one coarse search in the Y direction and one small two-dimensional fine search in both directions.
  • the total number of computations can be reduced to 2,600,000 additions/subtractions and 560 comparisons; representing a reduction of the total number of computations by roughly 115 times.
  • S1(i) is a quality factor of MSAD, and measures the difference between the value of MSAD and mean SAD.
  • MSAD quality increases with the value of S1(i), i.e., the deeper the SAD valley, the better MSAD quality.
  • the second parameter S2 (i) measures how much v(i) deviates from its past history.
  • the motion history 42 of the i'th block, h(i) (also shown in FIG. 2), is the recursively weighted averaging of previous final local motion vector outputs of the i'th block.
  • S2(i) is the vector difference between v(i) and h(i).
  • image motion is fairly smooth and consistent for both experienced and un-experienced ultrasound operators. If one value of v(i) has a very different direction and magnitude as compared with its history, it is very likely that this estimation is under the influence of noise or local tissue motion and does not accurately reflect true image local motion. In this case the v(i) estimation is not very reliable. Thus, a large value of S2(i) indicates that the estimated v(i) is less reliable.
  • a fuzzy logic control block 46 receives S1(i) and S2(i) as inputs, combines them using fuzzy rules (described below), and produces a single numerical output w(i) which represents a degree of accuracy of v(i).
  • the numerical output w(i) ranges from zero to one, with the estimation accuracy of v(i) increasing as w(i) approaches one.
  • the inputs S1(i) and S2(i) are first "fuzzified” into the linguistic expressions, or labels, "high”, “medium” and “low”.
  • Output w(i) also has its fuzzy expression as “very high”, “high”, “medium”, “low” and “very low”.
  • Membership functions of S1(i), S2(i) and w(i) are defined from a large number of experimental results, and are illustrated at FIGS. 7A through 7C, respectively.
  • the membership function of S1(i) is graphically illustrated at FIG. 7A as comprising three regions labeled as L (low), M (medium) and H (high). The regions overlap to a certain extent; specifically, the L and M regions overlap, and the M and H regions overlap.
  • the horizontal axis of the membership function graph defines the measured value of S1(i), and the vertical axis defines the degree of membership of the measured value within the defined label.
  • the membership function of S2(i) is graphically illustrated at FIG. 7B, and is constructed similar to the membership function of S1(i).
  • the membership function of w(i) is graphically illustrated at FIG. 7C, and is constructed similar to the membership functions of S1(i) and S2(i), though it includes five overlapping regions labeled as VL (very low), L (low), M (medium), H (high), and VH (very high).
  • fuzzy rules are used to define the relationship between S1(i), S2(i) and w(i). These fuzzy rules include:
  • the fuzzy rules are applied in parallel to determine the truth of the rules. For example, assume that measured values of S1(i) and S2(i) are 0.3 and 0.1, respectively. In FIG. 7A, a measured value of 0.3 relates to degrees of membership of approximately 0.65 in the L label and approximately 0.25 in the M label. In FIG. 7B, a measured value of 0.1 relates to a degree of membership of approximately 0.75 in the L label only. As a result, only the first two fuzzy rules are true, though they yield inconsistent results in that the first fuzzy rule concludes w(i) is medium and the second fuzzy rule concludes that w(i) is high. The output w(i) must be converted back to a numerical value, and the inconsistent results must be reconciled.
  • the low value of S1(i) is combined using a logical AND with the low value of S2(i) to provide the medium value of w(i).
  • the minimum value of the truth of the expressions is taken as the truth level of the rule.
  • the 0.65 degree of membership of S1(i) is less than the 0.75 degree of membership of S2(i), and is thus taken as the truth level for the first fuzzy rule.
  • the medium value of S1(i) is combined using a logical AND with the low value of S2(i) to provide the high value of w(i).
  • the 0.25 degree of membership of S1(i) is less than the 0.75 degree of membership of S2(i), and is thus taken as the truth level for the second fuzzy rule.
  • the M and H labels of the w(i) membership function are then truncated at the truth levels defined by the fuzzy rules, as illustrated graphically in FIG. 7D.
  • centroid defuzzification technique is used to convert the fuzzy output back to a numerical number w(i).
  • an estimate of the center of gravity is provided for the entire region determined to be true (illustrated as the shaded region of FIG. 7D). From FIG. 7D, the center of gravity of the shaded region is approximately 0.6, providing a numerical value for w(i).
  • the motion history h(i) is also updated recursively by weighting 1 mv(i).
  • L-S Minimum least-squares
  • the weighted L-S gives the local motion vector v(i) which has a larger w(i) more weight than those having smaller w(i). This way, the more reliable local motion vectors v(i) contribute more heavily to the optimization process.
  • the geometrically corrected frames are combined together at 54 of FIG. 2 to form a compound image.
  • the above method and apparatus for generating a compound image is applicable to both real-time imaging and re-generation of recorded image information.
  • a physician may use a conventional ultrasonic imaging system to produce image frames that are recorded onto a permanent storage medium, such as tape. Subsequently, the image frames could be processed into a compound image for later viewing by the physician by bringing the recorded image frame data to a viewing station. The viewing station would then process the image frame data by use of the method described above to generate a compound image.
  • the method and apparatus of this application are not limited to processing of ultrasonic images, but would be equally applicable to other imaging modalities, such as radar or photographic imaging.

Abstract

An image registration method and apparatus is provided in which consecutive ultrasonic image frames are correlated in order to derive transducer motion information. Since ultrasonic images from real-time scanning operations are highly correlated from one frame to another, the individual images can be divided into several smaller sub-image regions and a very fast and robust image motion detection algorithm utilized to measure the sub-image motion. This algorithm uses a fast adaptive coarse/fine minimum-sum-absolute-difference (MSAD) search strategy to compute an initial estimation of the local motion vectors. The initial estimation is combined with two measurement parameters by use of a fuzzy logic technique to derive a final estimation of the local motion vectors. The final local motion vectors are applied to a least-squares (L-S) process in order to estimate global image motion. Finally, a fast display technique generates compound images based on the derived global image motion.

Description

BACKGROUND OF THE INVENTION
1. Field of the Invention
This invention relates to ultrasonic diagnostic imaging, and more particularly, to an ultrasonic imaging system that produces a large compound image by correlating movement between consecutive image frames.
2. Description of Related Art
Ultrasonic imaging techniques are commonly used to produce two-dimensional diagnostic images of internal features of an object, such as a human anatomy. A diagnostic ultrasonic imaging system for medical use forms images of internal tissues of a human body by electrically exciting an acoustic transducer element or an array of acoustic transducer elements to generate short ultrasonic pulses that travel into the body. The ultrasonic pulses produce echoes as they reflect off of body tissues that appear as discontinuities or impedance changes to the propagating ultrasonic pulses. These echoes return to the transducer, and are converted back into electrical signals that are amplified and decoded to produce a cross-sectional image of the tissues. These ultrasonic imaging systems are of significant importance to the medical field by providing physicians with real-time, high resolution images of the internal features of a human anatomy without resort to more invasive exploratory techniques, such as surgery.
The acoustic transducer which radiates the ultrasonic pulses typically comprises a piezoelectric element or matrix of piezoelectric elements. As known in the art, a piezoelectric element deforms upon application of an electrical signal to produce the ultrasonic pulses. In a similar manner, the received echoes cause the piezoelectric element to deform and generate the corresponding electrical signal. The acoustic transducer is often packaged within a handheld device that allows the physician substantial freedom to easily manipulate the transducer over a desired area of interest. The transducer can then be electrically connected via a cable to a central control device that generates and processes the electrical signals. In turn, the control device transmits the image information to a real-time viewing device, such as a video display terminal. The image information may also be stored so that other physicians may view the diagnostic images at a later date.
The individual images produced by such ultrasonic imaging systems comprise discrete frames, with each such frame having a field of view limited by the relatively narrow region traversed by the ultrasonic pulses. As the transducer is manipulated along the body surface to obtain images of an adjacent region in the anatomy, each previous image is replaced on the viewing device by a new image defined by the limited field of view of the transducer. While a skilled physician can usually interpret the discrete frames in order to obtain a clear mental picture of the entire region traversed by the transducer, the discrete frames cannot be easily tiled together to produce a single, contiguous image. This can represent a significant drawback of conventional ultrasonic imaging systems, since it is not always possible for the physician to fully appreciate the overall condition of the body by consideration of the discrete frames alone. In some extreme cases, important information concerning the condition of the body tissues can be overlooked with serious potential consequences for the patient.
Previously, it has been demonstrated that a real-time compound ultrasonic image could be generated by use of so-called compound B-scanners. These B-scanners utilize a transducer mounted on an arm assembly that constrains the transducer to move along a single plane or axis. Either the arm assembly or the transducer element itself can be provided with sensing devices that track the precise position of the transducer. This positional information could then be utilized to register each of the discrete image frames together into a single composite image. An example of a compound B-scanner utilizing angular sensing devices on an arm assembly is disclosed in U.S. Pat. No. 4,431,007, to Amazeen et al., for REFERENCED REAL-TIME ULTRASONIC IMAGE DISPLAY. Despite this potential improvement in the art, conventional compound B-scanners are awkward and inflexible to operate due primarily to the relatively bulky mechanical arm assembly. Moreover, the sensing devices add significant complexity and cost to the ultrasonic imaging system. It should be apparent that the application of such prior art techniques to modern handheld ultrasonic transducers would be completely impractical in view of these significant drawbacks.
Thus, a critical need exists for a method to combine each of the discrete frames produced by an ultrasonic imaging system into a single image. The method should be compatible with modern handheld ultrasonic transducers without encumbering the handheld transducers with position sensing devices that increase the cost, weight and complexity of such imaging systems.
SUMMARY OF THE INVENTION
This application provides a method and apparatus for generating a large compound ultrasonic image. The method and apparatus utilizes image registration to eliminate the need for positional sensing devices altogether. In the image registration method and apparatus, consecutive moving image frames are correlated in order to derive the transducer motion information for compound image generation.
In particular, the method and apparatus recognizes that ultrasonic images from real-time scanning operations are highly correlated from one frame to another. Based on this recognition, the individual images are divided into several smaller sub-image regions, and a very fast and robust image motion detection algorithm is utilized to measure the sub-image motion. This algorithm uses a fast adaptive coarse/fine minimum-sum-absolute-difference (MSAD) search strategy to compute an initial estimation of the local motion vectors. The initial estimation is combined with two measurement parameters by use of a fuzzy logic technique to derive a final estimation of the local motion vectors. The final local motion vectors are applied to a least-squares (L-S) process in order to estimate global image motion. Finally, a fast display technique generates compound images based on the derived global image motion.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a partial perspective view of an ultrasonic imaging system adapted for real-time display of an exemplary fetal anatomy within a clinical environment;
FIG. 2 is a block diagram illustrating an image registration method in accordance with this invention;
FIG. 3 is a block diagram illustrating a minimum-sum-absolute-difference search of the image registration method;
FIG. 4 is a block diagram illustrating an alternative embodiment of the minimum-sum-absolute-difference search;
FIG. 5 is a three-dimensional graph illustrating derivation of a quality factor for the minimum-sum-absolute-difference search;
FIG. 6 is a block diagram illustrating derivation of a local vector deviation factor for the minimum-sum-absolute-difference search; and
FIGS. 7A through 7D are graphs illustrating fuzzy logic membership functions for the quality factor and the deviation factor to derive a control value and a single numerical output.
DETAILED DESCRIPTION
This invention satisfies the critical need for a method and apparatus to combine each of the discrete image frames produced by an ultrasonic imaging system into a single composite image. In the image registration method and apparatus of this invention, consecutive image frames are correlated in order to derive the transducer motion information for compound imaging. Significantly, the motion information is derived without encumbering the handheld transducer with position sensing devices that would otherwise increase the cost, weight and complexity of such ultrasonic imaging systems. In the detailed description that follows, like element numerals are used to describe like elements illustrated in one or more of the figures.
Referring first to FIG. 1, a prior art ultrasonic imaging system adapted for real-time display within a clinical environment is illustrated. A patient 10 carrying a fetus 12 is lying in a supine position to permit a physician or other technician 14 to perform an ultrasonic imaging operation. The imaging system comprises a scanhead 16 coupled to an ultrasound scanner 18. To perform the imaging operation, the physician 14 draws the scanhead 16 along the patient's skin surface in a direction which is parallel with the ultrasonic scan plane. An acoustically conductive lubricating agent may be applied to the skin prior to bringing the scanhead 16 into contact with the skin so as to improve acoustic coupling between the scanhead and the patient.
The scanhead 16 includes an ultrasonic transducer disposed at a surface thereof, comprising a piezoelectric element or matrix of individual piezoelectric elements. The ultrasound scanner 18 provides electrical signals to the scanhead 16 that cause it to generate ultrasonic pulses. The ultrasonic pulses 24 propagate in the scan plane through the skin of the patient 10 and echo off of the anatomical features of the fetus 12. The echoes return through the skin to the scanhead 16, which converts the echoes back into electrical signals received by the ultrasound scanner 18. The received electrical signals are then transmitted from the ultrasound scanner 18 to an image processor 20. The image processor 20 decodes the electrical signals into a two-dimensional, cross-sectional image that is displayed on the video display terminal 22. The image information may also be electronically stored in a memory medium, including a permanent storage medium such as a disk or tape drive, a temporary storage medium such as a solid state memory, or can be printed to a hard copy image such as a photograph.
The ultrasonic image displayed on the video display terminal 22 (or stored by use of the various other storage media described above) comprises only the information representative of the relatively narrow region traversed by the ultrasonic pulses 24. As illustrated in FIG. 1, only a portion of the anatomical features of the fetus 12 are displayed on the video display terminal 22. The region outside the sector of the ultrasonic pulses 24 (illustrated in phantom in FIG. 1) actually appears blank on the display device 22. As noted above, it would be very desirable to provide a large compound image, such as a single image that includes the entire image region 26.
Referring now to FIG. 2, a block diagram of an image registration method in accordance with this invention is illustrated. Two consecutive image frames 31, 32 are denoted as frame n-1 (the previous frame) and frame n (the current frame), respectively. The previous image frame n-1 is treated as a reference image. The current image frame n acquired by moving the transducer 16 of FIG. 1 along the skin surface is compared with the previous image frame n-1 to estimate motion of the transducer. In order to estimate image motion from frame n-1 to frame n, frame n is divided into a plurality of sub-frame regions 34, and the local motion of each of the sub-frame regions is estimated. For I total blocks, the i'th block of frame n is defined as n(i).
The particular size and number of the blocks 34 are selected by consideration of various factors. The block size should approximate the size of the particular image features. For example, image features such as tissue boundaries and blood vessels would require a relatively small block size. Nevertheless, such a small block size would not be suitable for estimation of motion because the small block size decorrelates very rapidly with relatively large magnitudes of motion. Moreover, the small block size would not be stable for in-vivo images where small scale tissue motion exists. At the same time, if the block size is too large, there will be too few blocks per image frame for motion estimation and the motion estimation will be unstable. Further, a large block size may introduce an image rotation error into the local motion estimation where the block is incorrectly assumed to have a translation component but no rotation component.
In a preferred embodiment of the invention, a block size of approximately 48×48 pixels is deemed acceptable for motion estimation in view of these factors in relation to an image frame size of about 200×400 pixels. A minimum block size of 32×32 pixels could also be utilized in relation to a smaller image size. These block sizes result in a block number I between approximately eight and twenty in order to yield the best results.
As illustrated in FIG. 3, local motion of each block n(i) of image n is estimated by moving the block n(i) (such as exemplary block 34) around on the n-1 image frame to find the best match. The search region size 33 on the n-1 image frame is less than the total area of the frame 31, and should relate to the expected image motion speed and image frame rate. For example, a large search region size 33 should be used when the transducer speed is high and also when the image frame rate is low. In a preferred embodiment, a search region size of 32 pixels in each of the four searching directions is adequate for use with an input image frame rate selected at 7.5 frame/second (computing one out of every four video frames), resulting in a search region size of 64×64 pixels.
The "best match" is determined by use of a minimum-sum-absolute-difference technique (MSAD) to match the n(i) block 34 with the search region 33 of frame n-1. The sum-absolute-difference (SAD) is the sum of absolute differences between corresponding pixel values between each block 34 and the search region 33. The "best match" between the block 34 and the search region 33 occurs where the SAD value is at a minimum. FIG. 5 illustrates a SAD map having values that extend in the X and Y directions. The generally uniform SAD map contains a valley where the SAD values dip below the mean, representing the location of the MSAD. The MSAD technique is used to derive a first estimation of local motion vector v(i) having a direction and magnitude indicating how the block n(i) translated from the n-1 image frame to the n image frame. The direction of the local motion vector v(i) is denoted by the arrow of FIG. 3.
It should be apparent that calculating the MSAD could be a very slow computing process if the number of blocks is large, the block size is large, and the search region is large. For example, a frame n having sixteen blocks 34 with a 48×48 pixel block size and a search region of 64×64 pixels would require about 300,000,000 separate additions/subtractions and about 65,000 comparisons to complete a single MSAD computation. Even with the relatively high speed of conventional processors, this would still encompass too much computation to accommodate real-time application. Accordingly, various search techniques are utilized in order to reduce the magnitude of the MSAD computation.
One such technique for calculating the MSAD utilizes a conventional MPEG encoder. MPEG, or Moving Picture Expert Group, is an industry accepted data compression standard for digitizing graphical information. MPEG encoders are commercially available that can be used to perform a rough estimate of MSAD location based on historical movement of the image. A "best match" is sought between an image characteristic within a block 34 and a search region by searching within a local neighborhood comprising a finite dimensional range, such as within ten pixels in the X and Y directions relative to the image characteristic. In FIG. 4, the location of the MSAD is identified by application of the MPEG encoder, and a local motion vector assigned to the block 34. A drawback of the use of an MPEG encoder is that it results in a high incidence of inaccurate local motion vectors, which must be filtered out.
The accuracy of the local motion vector determination can be further improved by consideration of certain unique constraints of compound ultrasonic imaging. Motion in the Y direction (vertical) from one frame to another is almost always smaller than the motion in the X direction (horizontal), because the ultrasonic transducer 16 generally moves only along the patient body contour. This transducer motion can be characterized as being more of a horizontal extension than an up/down folding. Further, the transducer motion is most likely to be in one direction and continuous; sudden reversals in motion direction are not very likely. It has been found from a large number of experiments that MSAD maps are often fairly smooth and continuous even for blocks which have a lot of noise and lack of image features.
In view of these constraints, a fast adaptive coarse/fine MSAD search strategy can be devised that significantly reduces the total amount of computation. The search strategy requires that the search range in the X direction be larger than that in the Y direction, and the search range and direction should be adaptive to the overall motion history. For example, if the history shows the image motion is in +X direction, it is very likely the subsequent frame will also move in the same direction. Because the MSAD map is smooth and continuous, a coarse search can be performed first to reduce the search region 33 to a smaller area, followed by a fine search within the smaller area. Because motion in the Y direction is usually very small (and usually zero), the two-dimensional MSAD search can be reduced to two one-dimensional searches in the X and Y directions, respectively. The first search should be conducted in the X direction to quickly narrow down the search region, followed by subsequent alternating one-dimensional searches in both the X and Y directions to quickly find the MSAD location. Points which have been searched during the coarse search or have been searched in the other direction could be skipped during the fine search. Based on the above search strategy, in most cases the location of the MSAD can be identified after one coarse and one medium search in the X direction, one coarse search in the Y direction and one small two-dimensional fine search in both directions. For the same example given above, the total number of computations can be reduced to 2,600,000 additions/subtractions and 560 comparisons; representing a reduction of the total number of computations by roughly 115 times.
Under the influence of image noise, tissue motion and other image artifacts, the first MSAD motion estimation is not always very reliable. Accordingly, two measures of the quality and reliability of the first local motion vector estimation v(i) are devised, termed S1(i) and S2(i). Referring to FIG. 5, S1(i) is a quality factor of MSAD, and measures the difference between the value of MSAD and mean SAD. MSAD quality increases with the value of S1(i), i.e., the deeper the SAD valley, the better MSAD quality. When strong image noise exists, or when there is a lack of image features, the SAD map will become more flat so that S1(i) becomes smaller. In that case, the estimation of v(i) becomes less reliable.
Referring to FIG. 6, the second parameter S2 (i) measures how much v(i) deviates from its past history. The motion history 42 of the i'th block, h(i) (also shown in FIG. 2), is the recursively weighted averaging of previous final local motion vector outputs of the i'th block. S2(i) is the vector difference between v(i) and h(i). Generally, image motion is fairly smooth and consistent for both experienced and un-experienced ultrasound operators. If one value of v(i) has a very different direction and magnitude as compared with its history, it is very likely that this estimation is under the influence of noise or local tissue motion and does not accurately reflect true image local motion. In this case the v(i) estimation is not very reliable. Thus, a large value of S2(i) indicates that the estimated v(i) is less reliable.
While the above estimation quality control concept is easy to understand, in practice it is difficult to implement since image changes caused by transducer motion can be rather complex. Accordingly, the quality and reliability of the motion estimate can be efficiently quantized by use of fuzzy logic. Referring back to FIG. 2, a fuzzy logic control block 46 receives S1(i) and S2(i) as inputs, combines them using fuzzy rules (described below), and produces a single numerical output w(i) which represents a degree of accuracy of v(i). The numerical output w(i) ranges from zero to one, with the estimation accuracy of v(i) increasing as w(i) approaches one.
The inputs S1(i) and S2(i) are first "fuzzified" into the linguistic expressions, or labels, "high", "medium" and "low". Output w(i) also has its fuzzy expression as "very high", "high", "medium", "low" and "very low". Membership functions of S1(i), S2(i) and w(i) are defined from a large number of experimental results, and are illustrated at FIGS. 7A through 7C, respectively. The membership function of S1(i) is graphically illustrated at FIG. 7A as comprising three regions labeled as L (low), M (medium) and H (high). The regions overlap to a certain extent; specifically, the L and M regions overlap, and the M and H regions overlap. The horizontal axis of the membership function graph defines the measured value of S1(i), and the vertical axis defines the degree of membership of the measured value within the defined label.
The membership function of S2(i) is graphically illustrated at FIG. 7B, and is constructed similar to the membership function of S1(i). Similarly, the membership function of w(i) is graphically illustrated at FIG. 7C, and is constructed similar to the membership functions of S1(i) and S2(i), though it includes five overlapping regions labeled as VL (very low), L (low), M (medium), H (high), and VH (very high).
Seven fuzzy rules are used to define the relationship between S1(i), S2(i) and w(i). These fuzzy rules include:
(1) If S1(i) is low (L) AND S2(i) is also low (L) , then w(i) is medium (M);
(2) If S1(i) is medium (M) AND S2(i) is low (L), then w(i) is high (H);
(3) If S1(i) is high (H) AND S2(i) is low (L), then w(i) is very high (VH);
(4) If S1(i) is low (L) AND S2(i) is medium (M), the w(i) is low (L);
(5) If S1(i) is medium (M) AND S2(i) is also medium (M), then w(i) is medium (M);
(6) If S1(i) is high (H) AND S2(i) is medium (M), then w(i) is high (H); and
(7) If S2(i) is high (H), then w(i) is very low (VL).
The fuzzy rules are applied in parallel to determine the truth of the rules. For example, assume that measured values of S1(i) and S2(i) are 0.3 and 0.1, respectively. In FIG. 7A, a measured value of 0.3 relates to degrees of membership of approximately 0.65 in the L label and approximately 0.25 in the M label. In FIG. 7B, a measured value of 0.1 relates to a degree of membership of approximately 0.75 in the L label only. As a result, only the first two fuzzy rules are true, though they yield inconsistent results in that the first fuzzy rule concludes w(i) is medium and the second fuzzy rule concludes that w(i) is high. The output w(i) must be converted back to a numerical value, and the inconsistent results must be reconciled.
Under the first fuzzy rule, the low value of S1(i) is combined using a logical AND with the low value of S2(i) to provide the medium value of w(i). Under the logical AND operation, the minimum value of the truth of the expressions is taken as the truth level of the rule. In other words, the 0.65 degree of membership of S1(i) is less than the 0.75 degree of membership of S2(i), and is thus taken as the truth level for the first fuzzy rule. Similarly, under the second fuzzy rule, the medium value of S1(i) is combined using a logical AND with the low value of S2(i) to provide the high value of w(i). The 0.25 degree of membership of S1(i) is less than the 0.75 degree of membership of S2(i), and is thus taken as the truth level for the second fuzzy rule. The M and H labels of the w(i) membership function are then truncated at the truth levels defined by the fuzzy rules, as illustrated graphically in FIG. 7D.
Finally, a centroid defuzzification technique is used to convert the fuzzy output back to a numerical number w(i). Using this technique, an estimate of the center of gravity is provided for the entire region determined to be true (illustrated as the shaded region of FIG. 7D). From FIG. 7D, the center of gravity of the shaded region is approximately 0.6, providing a numerical value for w(i).
After the reliability parameter w(i) is obtained, the next step is to use w(i) to improve the local motion estimation v(i). If w(i) is large, v(i) is used directly as the final local motion vector 1 mv(i). In contrast, if w(i) is very small, the average frame motion history h(i) is used as the estimated 1 mv(i), since the motion history is more likely to be a better estimation than the less reliable v(i). If w(i) is neither very large nor very small, it is used as a weighing factor to average v(i) and h(i). For instance, if w(i)=0.6, as in the above example, then 1 mv(i)=0.6*v(i)+(1-0.6)*h(i).
The motion history h(i) is also updated recursively by weighting 1 mv(i). The weight is selected to be between zero and one; a larger weight value makes the very recent motion estimation contribute more to the history h(i). For example, if the weight is 0.5, then h(i)=0.5*1 mv(i)+(1-0.5)*h(i).
Once all of the local motion vector outputs 1 mv(i) are estimated for the n'th frame, the outputs are combined together at 48 of FIG. 2 to estimate the frame global motion vector, gmv(i). Minimum least-squares (L-S) error parameter fitting is used to combine the motion vector outputs, using three optimization parameters including frame translation (Xn, Yn) and rotation en. The weighted L-S gives the local motion vector v(i) which has a larger w(i) more weight than those having smaller w(i). This way, the more reliable local motion vectors v(i) contribute more heavily to the optimization process. The geometrically corrected frames are combined together at 54 of FIG. 2 to form a compound image. Three techniques can be applied to combine the corrected frames, including "image growing" which only puts new pixel data in the non-overlapping part to the compound image buffer, "recursive spatial compounding" which recursively averages the new image frame with existing compound image, and "ramp compounding" which gives weight ramps for both new image frame and existing compound image in the overlapping area. This latter technique successfully reduces compound image local discontinuity caused by motion jitters. Finally, the compound image is displayed on a video display terminal 56 or other such device, providing the full image represented in phantom on the terminal 22 of FIG. 1, described above.
It should be apparent that the above method and apparatus for generating a compound image is applicable to both real-time imaging and re-generation of recorded image information. In application, a physician may use a conventional ultrasonic imaging system to produce image frames that are recorded onto a permanent storage medium, such as tape. Subsequently, the image frames could be processed into a compound image for later viewing by the physician by bringing the recorded image frame data to a viewing station. The viewing station would then process the image frame data by use of the method described above to generate a compound image. It should also be apparent that the method and apparatus of this application are not limited to processing of ultrasonic images, but would be equally applicable to other imaging modalities, such as radar or photographic imaging.

Claims (24)

What is claimed is:
1. A method for providing a compound image, comprising the steps of:
generating a plurality of substantially adjacent image frames of an interrogation region of a patient's body;
dividing individual ones of the image frames into a plurality of sub-image regions;
estimating local motion vectors of the respective sub-image regions between consecutive ones of the image frames;
estimating global image motion based on the estimated local motion vectors; and
displaying a compound image based on the estimated global image motion.
2. The method of claim 1, wherein the step of dividing the individual image frames into a plurality of sub-image regions further comprises the step of dividing the individual image frames into a range of approximately eight to twenty sub-image regions.
3. The method of claim 1, wherein the estimating steps are performed concurrently with the generating step so that a real-time compound image is provided in the displaying step.
4. A method for providing a compound image from a plurality of individual image frames, comprising the steps of:
dividing the individual image frames into a plurality of sub-image regions;
estimating local motion vectors of the respective sub-image regions between consecutive ones of the individual image frames;
estimating global image motion based on the estimated local motion vectors; and
displaying a compound image based on the estimated global image motion;
wherein the step of estimating local motion vectors further comprises the steps of:
calculating an initial estimation of the local motion vectors using a fast adaptive coarse/fine minimum-sum-absolute-difference (MSAD) search; and
deriving a final estimation of the local motion vectors from the initial estimation of the local motion vectors by use of a fuzzy logic technique.
5. The method of claim 4, wherein the step of deriving a final estimation further comprises the step of defining a first quality factor of the MSAD by measuring the difference between the value of MSAD and mean sum-absolute-difference (SAD).
6. The method of claim 5, wherein the step of deriving a final estimation further comprises the step of defining a second quality factor of MSAD by measuring deviation of the initial estimation of the local motion vector from an historical measurement of the local motion vector.
7. The method of claim 6, wherein the step of deriving a final estimation further comprises the step of defining membership functions for the first and second quality factors.
8. The method of claim 7, wherein the step of deriving a final estimation further comprises the step of determining a membership value for the first and second quality factors.
9. The method of claim 8, wherein the step of deriving a final estimation further comprises the step of combining the membership values for the first and second quality factors in accordance with predefined fuzzy rules to generate an output value.
10. The method of claim 9, wherein the step of deriving a final estimation further comprises the step of averaging the initial estimation of the local motion vectors with the historical estimation of the local motion vectors using the output value as a weight.
11. The method of claim 4, wherein the step of calculating an initial estimation of the local motion vectors further comprises the steps of conducting a one-dimensional search in a first axial direction and at least one additional one-dimensional search in a second axial direction perpendicular to said first axial direction.
12. The method of claim 4, wherein the step of calculating an initial estimation of the local motion vectors further comprises the steps of defining a search region having an area greater than an associated one of the sub-image regions, the one-dimensional searches being conducted within the search region.
13. A method for providing a compound image from a plurality of individual image frames, comprising the steps of:
dividing the individual image frames into a plurality of sub-image regions;
estimating local motion vectors of the respective sub-image regions between consecutive ones of the individual image frames;
estimating global image motion based on the estimated local motion vectors; and
displaying a compound image based on the estimated global image motion;
wherein the step of estimating global image motion further comprises the step of applying the estimated local motion vectors to a least-squares (L-S) process.
14. An apparatus for providing a compound image, comprising:
an ultrasonic scanner providing a plurality of individual ultrasonic image frames;
means for dividing the individual image frames into a plurality of sub-image regions;
means for estimating local motion vectors of the respective sub-image regions between consecutive ones of the individual image frames by use of a minimum-sum-absolute-difference search;
means for estimating global image motion based on the estimated local motion vectors; and
means for displaying a compound image based on the estimated global image motion.
15. The apparatus of claim 14, wherein the means for estimating local motion vectors further comprises:
means for calculating an initial estimation of the local motion vectors using a fast adaptive coarse/fine minimum-sum-absolute-difference (MSAD) search; and
means for deriving a final estimation of the local motion vectors from the initial estimation of the local motion vectors by use of a fuzzy logic technique.
16. The apparatus of claim 15, wherein the deriving means further comprises means for defining a first quality factor of the MSAD by measuring the difference between the value of MSAD and mean sum-absolute-difference (SAD).
17. The apparatus of claim 16, wherein the deriving means further comprises means for defining a second quality factor of MSAD by measuring deviation of the initial estimation of the local motion vector from an historical measurement of the local motion vector.
18. The apparatus of claim 17, wherein the deriving means further comprises means for defining membership functions for the first and second quality factors.
19. The apparatus of claim 18, wherein the deriving means further comprises means for determining a membership value for the first and second quality factors.
20. The apparatus of claim 19, wherein the deriving means further comprises means for combining the membership values for the first and second quality factors in accordance with predefined fuzzy rules to generate an output value.
21. The apparatus of claim 20, wherein the deriving means further comprises means for averaging the initial estimation of the local motion vectors with the historical estimation of the local motion vectors using the output value as a weight.
22. An apparatus for providing a compound ultrasonic image, comprising:
an ultrasonic scanner providing a plurality of substantially adjacent ultrasonic image frames, said image frames each respectively having a plurality of sub-image regions;
means for estimating local motion of the respective sub-image regions between consecutive ones of the image frames by calculating an initial estimation of the local motion and deriving a final estimation of the local motion from the initial estimation;
means for estimating global image motion based on the estimated local motion; and
means for displaying a compound image based on the estimated global image motion.
23. The apparatus of claim 22, wherein said means for estimating local image motion further comprises a fast adaptive coarse/fine minimum-sum-absolute-difference (MSAD) search.
24. The apparatus of claim 22, wherein both said means for estimating local and global image motion are adapted to operate concurrently with said ultrasonic scanner to permit said displaying means to provide a real-time compound image.
US08/414,978 1995-03-31 1995-03-31 Method and apparatus for generating large compound ultrasound image Expired - Lifetime US5575286A (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
US08/414,978 US5575286A (en) 1995-03-31 1995-03-31 Method and apparatus for generating large compound ultrasound image
DE19611990A DE19611990C2 (en) 1995-03-31 1996-03-26 Method and device for producing large, composite ultrasound images
JP07924396A JP3345257B2 (en) 1995-03-31 1996-04-01 Apparatus for generating an ultrasound image of a composite, enlarged field of view
US08/728,280 US5899861A (en) 1995-03-31 1996-10-08 3-dimensional volume by aggregating ultrasound fields of view
US08/747,429 US5782766A (en) 1995-03-31 1996-11-19 Method and apparatus for generating and displaying panoramic ultrasound images

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US08/414,978 US5575286A (en) 1995-03-31 1995-03-31 Method and apparatus for generating large compound ultrasound image

Related Child Applications (2)

Application Number Title Priority Date Filing Date
US08/728,280 Continuation-In-Part US5899861A (en) 1995-03-31 1996-10-08 3-dimensional volume by aggregating ultrasound fields of view
US08/747,429 Continuation-In-Part US5782766A (en) 1995-03-31 1996-11-19 Method and apparatus for generating and displaying panoramic ultrasound images

Publications (1)

Publication Number Publication Date
US5575286A true US5575286A (en) 1996-11-19

Family

ID=23643828

Family Applications (1)

Application Number Title Priority Date Filing Date
US08/414,978 Expired - Lifetime US5575286A (en) 1995-03-31 1995-03-31 Method and apparatus for generating large compound ultrasound image

Country Status (3)

Country Link
US (1) US5575286A (en)
JP (1) JP3345257B2 (en)
DE (1) DE19611990C2 (en)

Cited By (118)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5782766A (en) * 1995-03-31 1998-07-21 Siemens Medical Systems, Inc. Method and apparatus for generating and displaying panoramic ultrasound images
US5876345A (en) * 1997-02-27 1999-03-02 Acuson Corporation Ultrasonic catheter, system and method for two dimensional imaging or three-dimensional reconstruction
US5876342A (en) * 1997-06-30 1999-03-02 Siemens Medical Systems, Inc. System and method for 3-D ultrasound imaging and motion estimation
US5899861A (en) * 1995-03-31 1999-05-04 Siemens Medical Systems, Inc. 3-dimensional volume by aggregating ultrasound fields of view
US5956418A (en) * 1996-12-10 1999-09-21 Medsim Ltd. Method of mosaicing ultrasonic volumes for visual simulation
US6014473A (en) 1996-02-29 2000-01-11 Acuson Corporation Multiple ultrasound image registration system, method and transducer
US6045508A (en) * 1997-02-27 2000-04-04 Acuson Corporation Ultrasonic probe, system and method for two-dimensional imaging or three-dimensional reconstruction
WO2000020887A1 (en) * 1998-10-01 2000-04-13 Koninklijke Philips Electronics N.V. Ultrasonic diagnostic imaging system with adaptive spatial compounding
US6101238A (en) * 1998-11-25 2000-08-08 Siemens Corporate Research, Inc. System for generating a compound x-ray image for diagnosis
US6117081A (en) * 1998-10-01 2000-09-12 Atl Ultrasound, Inc. Method for correcting blurring of spatially compounded ultrasonic diagnostic images
US6116244A (en) * 1998-06-02 2000-09-12 Acuson Corporation Ultrasonic system and method for three-dimensional imaging with opacity control
US6152878A (en) * 1997-06-19 2000-11-28 Medinol Ltd. Intravascular ultrasound enhanced image and signal processing
US6159152A (en) * 1998-10-26 2000-12-12 Acuson Corporation Medical diagnostic ultrasound system and method for multiple image registration
US6193657B1 (en) * 1998-12-31 2001-02-27 Ge Medical Systems Global Technology Company, Llc Image based probe position and orientation detection
US6224552B1 (en) 1998-10-01 2001-05-01 Atl Ultrasound Ultrasonic diagnostic imaging system with reduced spatial compounding seam artifacts
US6234968B1 (en) 1999-06-15 2001-05-22 Acuson Corporation 3-D diagnostic medical ultrasound imaging using a 1-D array
US6238345B1 (en) 1999-06-30 2001-05-29 Atl Ultrasound Image memory for extended field of view ultrasonic diagnostic imaging
US6254539B1 (en) 1999-08-26 2001-07-03 Acuson Corporation Transducer motion compensation in medical diagnostic ultrasound 3-D imaging
US20010012385A1 (en) * 1998-12-30 2001-08-09 John A Hossack Medical diagnostic ultrasonic imaging method and system for displaying multi-phase, multi-frame images
US6283917B1 (en) 1998-10-01 2001-09-04 Atl Ultrasound Ultrasonic diagnostic imaging system with blurring corrected spatial compounding
US6299579B1 (en) 1999-06-30 2001-10-09 Atl Ultrasound Extended field of view ultrasonic diagnostic imaging with image reacquisition
US6328693B1 (en) 1999-01-28 2001-12-11 Hitachi, Ltd. Method and device of synthesizing panorama image from ultrasound images
US6364835B1 (en) 1998-11-20 2002-04-02 Acuson Corporation Medical diagnostic ultrasound imaging methods for extended field of view
US6390981B1 (en) 2000-05-23 2002-05-21 Koninklijke Philips Electronics N.V. Ultrasonic spatial compounding with curved array scanheads
US6416477B1 (en) 2000-08-22 2002-07-09 Koninklijke Philips Electronics N.V. Ultrasonic diagnostic systems with spatial compounded panoramic imaging
US6423004B1 (en) * 2000-05-30 2002-07-23 Ge Medical Systems Global Technology Company, Llc Real-time ultrasound spatial compounding using multiple angles of view
US6442289B1 (en) 1999-06-30 2002-08-27 Koninklijke Philips Electronics N.V. Extended field of view ultrasonic diagnostic imaging
US20020167533A1 (en) * 2001-04-25 2002-11-14 Siemens Medical Systems, Inc. Compound image display system and method
US6503201B1 (en) 2001-10-03 2003-01-07 Koninklijke Philips Electronics N.V. Correction of extended field of view images for distortion due to scanhead motion
US6544177B1 (en) 1998-10-01 2003-04-08 Atl Ultrasound, Inc. Ultrasonic diagnostic imaging system and method with harmonic spatial compounding
US6547732B2 (en) 1998-10-01 2003-04-15 Koninklijke Philips Electronics N.V. Adaptive image processing for spatial compounding
US6554770B1 (en) 1998-11-20 2003-04-29 Acuson Corporation Medical diagnostic ultrasound imaging methods for extended field of view
US6567469B1 (en) 2000-03-23 2003-05-20 Koninklijke Philips Electronics N.V. Motion estimation algorithm suitable for H.261 videoconferencing applications
US6572549B1 (en) 2001-12-18 2003-06-03 Koninklijke Philips Electronics Nv High frame rate extended field of view ultrasound imaging system and method
US6589176B2 (en) 2001-12-05 2003-07-08 Koninklijke Philips Electronics N.V. Ultrasonic image stabilization system and method
US6605042B2 (en) 2001-08-10 2003-08-12 Ge Medical Systems Global Technology Company, Llc Method and apparatus for rotation registration of extended field of view ultrasound images
US6685643B1 (en) * 1998-08-21 2004-02-03 Tomtec Imaging Systems Gmbh Method and device for recording ultrasonic images
US20040064036A1 (en) * 2002-09-26 2004-04-01 Zuhua Mao Methods and systems for motion tracking
US20040080467A1 (en) * 2002-10-28 2004-04-29 University Of Washington Virtual image registration in augmented display field
US6755787B2 (en) 1998-06-02 2004-06-29 Acuson Corporation Medical diagnostic ultrasound system and method for versatile processing
US20040127796A1 (en) * 2002-06-07 2004-07-01 Vikram Chalana 3D ultrasound-based instrument for non-invasive measurement of amniotic fluid volume
WO2004063990A1 (en) * 2003-01-13 2004-07-29 Koninklijke Philips Electronics N.V. A method of image registration and medical image data processing apparatus
US20050033173A1 (en) * 2003-08-05 2005-02-10 Von Behren Patrick L. Extended volume ultrasound data acquisition
US6858010B2 (en) 2003-05-06 2005-02-22 Siemens Medical Solutions Usa, Inc. Identifying clinical markers in spatial compounding ultrasound imaging
US20050096539A1 (en) * 2003-10-31 2005-05-05 Siemens Medical Solutions Usa, Inc. Intelligent ultrasound examination storage system
US20050259737A1 (en) * 2004-05-18 2005-11-24 Sunplus Technology Co., Ltd. Optimal correlation matching method and system for determining track behavior
US20060013501A1 (en) * 2004-07-09 2006-01-19 Tadashi Tamura Method and apparatus for image processing
US6996177B1 (en) 1999-08-02 2006-02-07 Koninklijke Philips Electronics N.V. Motion estimation
US20060034528A1 (en) * 2004-08-12 2006-02-16 Yeping Su System and method for non-iterative global motion estimation
US20060173292A1 (en) * 2002-09-12 2006-08-03 Hirotaka Baba Biological tissue motion trace method and image diagnosis device using the trace method
US20060228011A1 (en) * 2005-04-06 2006-10-12 Everett Matthew J Method and apparatus for measuring motion of a subject using a series of partial images from an imaging system
US20060235301A1 (en) * 2002-06-07 2006-10-19 Vikram Chalana 3D ultrasound-based instrument for non-invasive measurement of amniotic fluid volume
US20070211928A1 (en) * 2005-11-10 2007-09-13 Rosetta Inpharmatics Llc Discover biological features using composite images
US20070255137A1 (en) * 2006-05-01 2007-11-01 Siemens Medical Solutions Usa, Inc. Extended volume ultrasound data display and measurement
US20080021317A1 (en) * 2006-07-24 2008-01-24 Siemens Medical Solutions Usa, Inc. Ultrasound medical imaging with robotic assistance for volume imaging
US20080146932A1 (en) * 2002-06-07 2008-06-19 Vikram Chalana 3D ultrasound-based instrument for non-invasive measurement of Amniotic Fluid Volume
CN100455266C (en) * 2005-03-29 2009-01-28 深圳迈瑞生物医疗电子股份有限公司 Broad image processing method
US20100160783A1 (en) * 2008-12-19 2010-06-24 Menachem Halmann Ultrasound imaging method and apparatus
US7780601B2 (en) 2007-06-05 2010-08-24 Siemens Medical Solutions Usa, Inc. Adaptive clinical marker preservation in spatial compound ultrasound imaging
US7819806B2 (en) 2002-06-07 2010-10-26 Verathon Inc. System and method to identify and measure organ wall boundaries
US7837624B1 (en) * 1998-11-20 2010-11-23 Siemens Medical Solutions Usa, Inc. Medical diagnostic ultrasound imaging methods for extended field of view
US8133181B2 (en) 2007-05-16 2012-03-13 Verathon Inc. Device, system and method to measure abdominal aortic aneurysm diameter
US8167803B2 (en) 2007-05-16 2012-05-01 Verathon Inc. System and method for bladder detection using harmonic imaging
US8221322B2 (en) 2002-06-07 2012-07-17 Verathon Inc. Systems and methods to improve clarity in ultrasound images
US8221321B2 (en) 2002-06-07 2012-07-17 Verathon Inc. Systems and methods for quantification and classification of fluids in human cavities in ultrasound images
US20120212618A1 (en) * 2011-02-21 2012-08-23 Samsung Electronics Co., Ltd. Method of ultrasonic beamforming and apparatus therefor
US8308644B2 (en) 2002-08-09 2012-11-13 Verathon Inc. Instantaneous ultrasonic measurement of bladder volume
US20130044929A1 (en) * 2011-08-19 2013-02-21 Industrial Technology Research Institute Ultrasound image registration apparatus and method thereof
US8409103B2 (en) 2005-05-06 2013-04-02 Vasonova, Inc. Ultrasound methods of positioning guided vascular access devices in the venous system
US8597193B2 (en) 2005-05-06 2013-12-03 Vasonova, Inc. Apparatus and method for endovascular device guiding and positioning using physiological parameters
US8711366B2 (en) 2005-01-21 2014-04-29 Carl Zeiss Meditec, Inc. Method of motion correction in optical coherence tomography imaging
US8857988B2 (en) 2011-07-07 2014-10-14 Carl Zeiss Meditec, Inc. Data acquisition methods for reduced motion artifacts and applications in OCT angiography
US8891840B2 (en) 2012-02-13 2014-11-18 Siemens Medical Solutions Usa, Inc. Dynamic steered spatial compounding in ultrasound imaging
US8965490B2 (en) 2012-05-07 2015-02-24 Vasonova, Inc. Systems and methods for detection of the superior vena cava area
US20150063628A1 (en) * 2013-09-04 2015-03-05 Xerox Corporation Robust and computationally efficient video-based object tracking in regularized motion environments
US9033510B2 (en) 2011-03-30 2015-05-19 Carl Zeiss Meditec, Inc. Systems and methods for efficiently obtaining measurements of the human eye using tracking
US9081097B2 (en) 2012-05-01 2015-07-14 Siemens Medical Solutions Usa, Inc. Component frame enhancement for spatial compounding in ultrasound imaging
US9119551B2 (en) 2010-11-08 2015-09-01 Vasonova, Inc. Endovascular navigation system and method
US20150342560A1 (en) * 2013-01-25 2015-12-03 Ultrasafe Ultrasound Llc Novel Algorithms for Feature Detection and Hiding from Ultrasound Images
US9445780B2 (en) 2009-12-04 2016-09-20 University Of Virginia Patent Foundation Tracked ultrasound vessel imaging
US9706914B2 (en) 2012-01-19 2017-07-18 Carl Zeiss Meditec, Inc. Systems and methods for enhanced accuracy in OCT imaging of the cornea
WO2018046455A1 (en) 2016-09-09 2018-03-15 Koninklijke Philips N.V. Stabilization of ultrasound images
US10315222B2 (en) 2016-05-04 2019-06-11 Invensense, Inc. Two-dimensional array of CMOS control elements
US10325915B2 (en) 2016-05-04 2019-06-18 Invensense, Inc. Two-dimensional array of CMOS control elements
US10368837B2 (en) 2005-05-06 2019-08-06 Arrow International, Inc. Apparatus and method for vascular access
US10408797B2 (en) 2016-05-10 2019-09-10 Invensense, Inc. Sensing device with a temperature sensor
US10441975B2 (en) 2016-05-10 2019-10-15 Invensense, Inc. Supplemental sensor modes and systems for ultrasonic transducers
US10445547B2 (en) 2016-05-04 2019-10-15 Invensense, Inc. Device mountable packaging of ultrasonic transducers
US10452887B2 (en) 2016-05-10 2019-10-22 Invensense, Inc. Operating a fingerprint sensor comprised of ultrasonic transducers
US10474862B2 (en) * 2017-06-01 2019-11-12 Invensense, Inc. Image generation in an electronic device using ultrasonic transducers
US10539539B2 (en) 2016-05-10 2020-01-21 Invensense, Inc. Operation of an ultrasonic sensor
US10562070B2 (en) 2016-05-10 2020-02-18 Invensense, Inc. Receive operation of an ultrasonic sensor
US10600403B2 (en) 2016-05-10 2020-03-24 Invensense, Inc. Transmit operation of an ultrasonic sensor
US10632500B2 (en) 2016-05-10 2020-04-28 Invensense, Inc. Ultrasonic transducer with a non-uniform membrane
US10643052B2 (en) 2017-06-28 2020-05-05 Invensense, Inc. Image generation in an electronic device using ultrasonic transducers
US10656255B2 (en) 2016-05-04 2020-05-19 Invensense, Inc. Piezoelectric micromachined ultrasonic transducer (PMUT)
US10670716B2 (en) 2016-05-04 2020-06-02 Invensense, Inc. Operating a two-dimensional array of ultrasonic transducers
US10706835B2 (en) 2016-05-10 2020-07-07 Invensense, Inc. Transmit beamforming of a two-dimensional array of ultrasonic transducers
US10755067B2 (en) 2018-03-22 2020-08-25 Invensense, Inc. Operating a fingerprint sensor comprised of ultrasonic transducers
CN112012062A (en) * 2020-09-01 2020-12-01 天津动网信息科技有限公司 Rail and roadbed offset three-coordinate monitoring device and method
US10891461B2 (en) 2017-05-22 2021-01-12 Invensense, Inc. Live fingerprint detection utilizing an integrated ultrasound and infrared sensor
US10936843B2 (en) 2018-12-28 2021-03-02 Invensense, Inc. Segmented image acquisition
US10936841B2 (en) 2017-12-01 2021-03-02 Invensense, Inc. Darkfield tracking
US10984209B2 (en) 2017-12-01 2021-04-20 Invensense, Inc. Darkfield modeling
US10997388B2 (en) 2017-12-01 2021-05-04 Invensense, Inc. Darkfield contamination detection
US11151355B2 (en) 2018-01-24 2021-10-19 Invensense, Inc. Generation of an estimated fingerprint
US11176345B2 (en) 2019-07-17 2021-11-16 Invensense, Inc. Ultrasonic fingerprint sensor with a contact layer of non-uniform thickness
US11188735B2 (en) 2019-06-24 2021-11-30 Invensense, Inc. Fake finger detection using ridge features
US11185305B2 (en) * 2016-06-30 2021-11-30 Koninklijke Philips N.V. Intertial device tracking system and method of operation thereof
US11216681B2 (en) 2019-06-25 2022-01-04 Invensense, Inc. Fake finger detection based on transient features
US11216632B2 (en) 2019-07-17 2022-01-04 Invensense, Inc. Ultrasonic fingerprint sensor with a contact layer of non-uniform thickness
US11232549B2 (en) 2019-08-23 2022-01-25 Invensense, Inc. Adapting a quality threshold for a fingerprint image
US11243300B2 (en) 2020-03-10 2022-02-08 Invensense, Inc. Operating a fingerprint sensor comprised of ultrasonic transducers and a presence sensor
US11298109B2 (en) * 2013-03-11 2022-04-12 Canon Medical Systems Corporation Ultrasonic diagnostic apparatus and image processing apparatus
US11328165B2 (en) 2020-04-24 2022-05-10 Invensense, Inc. Pressure-based activation of fingerprint spoof detection
US11392789B2 (en) 2019-10-21 2022-07-19 Invensense, Inc. Fingerprint authentication using a synthetic enrollment image
US11460957B2 (en) 2020-03-09 2022-10-04 Invensense, Inc. Ultrasonic fingerprint sensor with a contact layer of non-uniform thickness
US11673165B2 (en) 2016-05-10 2023-06-13 Invensense, Inc. Ultrasonic transducer operable in a surface acoustic wave (SAW) mode

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3556129B2 (en) * 1999-07-09 2004-08-18 松下電器産業株式会社 Ultrasound diagnostic equipment
JP4659974B2 (en) * 2000-12-12 2011-03-30 株式会社東芝 Ultrasonic diagnostic equipment
JP4785105B2 (en) * 2001-08-03 2011-10-05 株式会社日立メディコ Ultrasonic image processing device
JP4958348B2 (en) * 2001-09-06 2012-06-20 株式会社日立メディコ Ultrasonic imaging device
JP4704630B2 (en) * 2001-09-14 2011-06-15 アロカ株式会社 Ultrasonic panoramic image forming device
JP2003093382A (en) * 2001-09-26 2003-04-02 Matsushita Electric Ind Co Ltd Ultrasonograph
KR100490564B1 (en) * 2002-07-23 2005-05-19 주식회사 메디슨 Apparatus and method for recognizing organ from ultrasound image signal
KR100869497B1 (en) * 2005-07-01 2008-11-21 주식회사 메디슨 Hierarchical Motion Estimating Method and Ultrasound Image Display Device Using the Same
JP5002397B2 (en) 2007-09-28 2012-08-15 株式会社東芝 Ultrasonic diagnostic apparatus and program
WO2009147620A2 (en) * 2008-06-05 2009-12-10 Koninklijke Philips Electronics, N.V. Extended field of view ultrasonic imaging with a two dimensional array probe
JP5271029B2 (en) * 2008-10-16 2013-08-21 株式会社東芝 Ultrasonic diagnostic equipment
JP4709937B2 (en) * 2010-10-01 2011-06-29 株式会社東芝 Ultrasonic diagnostic apparatus and image processing apparatus

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4431007A (en) * 1981-02-04 1984-02-14 General Electric Company Referenced real-time ultrasonic image display
US4580219A (en) * 1983-05-02 1986-04-01 General Electric Company Method for reducing image artifacts due to projection measurement inconsistencies
US5045939A (en) * 1989-07-06 1991-09-03 Matsushita Electric Industrial Co., Ltd. Apparatus utilizing motion detector for converting a maximum motion portion of a wide screen tv signal to a normal screen tv signal
US5051826A (en) * 1989-02-28 1991-09-24 Kabushiki Kaisha Toshiba Vertical edge detection circuit for a television image motion adaptive progressive scanning conversion circuit
US5107837A (en) * 1989-11-17 1992-04-28 Board Of Regents, University Of Texas Method and apparatus for measurement and imaging of tissue compressibility or compliance
US5247937A (en) * 1989-11-17 1993-09-28 Board Of Regents, The University Of Texas System Transaxial compression technique for sound velocity estimation
US5257626A (en) * 1990-11-26 1993-11-02 Board Of Trustees Of The Leland Stanford Junior University Method of noninvasive myocardial motion analysis using bidirectional motion intergration in phase contrast MRI maps of myocardial velocity
US5262856A (en) * 1992-06-04 1993-11-16 Massachusetts Institute Of Technology Video image compositing techniques
US5293870A (en) * 1989-11-17 1994-03-15 Board Of Regents The University Of Texas System Method and apparatus for elastographic measurement and imaging
US5305104A (en) * 1992-07-27 1994-04-19 The Trustees Of Columbia University In The City Of New York Digitally assisted motion compensated deinterlacing for enhanced definition television
US5353354A (en) * 1990-11-22 1994-10-04 Advanced Technology Laboratories, Inc. Acquisition and display of ultrasonic images from sequentially oriented image planes
US5474070A (en) * 1989-11-17 1995-12-12 The Board Of Regents Of The University Of Texas System Method and apparatus for elastographic measurement and imaging

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CA1242267A (en) * 1984-09-25 1988-09-20 Rainer Fehr Real time display of an ultrasonic compound image
JPH0528243A (en) * 1991-07-24 1993-02-05 Toshiba Corp Image-forming device

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4431007A (en) * 1981-02-04 1984-02-14 General Electric Company Referenced real-time ultrasonic image display
US4580219A (en) * 1983-05-02 1986-04-01 General Electric Company Method for reducing image artifacts due to projection measurement inconsistencies
US5051826A (en) * 1989-02-28 1991-09-24 Kabushiki Kaisha Toshiba Vertical edge detection circuit for a television image motion adaptive progressive scanning conversion circuit
US5045939A (en) * 1989-07-06 1991-09-03 Matsushita Electric Industrial Co., Ltd. Apparatus utilizing motion detector for converting a maximum motion portion of a wide screen tv signal to a normal screen tv signal
US5247937A (en) * 1989-11-17 1993-09-28 Board Of Regents, The University Of Texas System Transaxial compression technique for sound velocity estimation
US5178147A (en) * 1989-11-17 1993-01-12 Board Of Regents, The University Of Texas System Method and apparatus for elastographic measurement and imaging
US5107837A (en) * 1989-11-17 1992-04-28 Board Of Regents, University Of Texas Method and apparatus for measurement and imaging of tissue compressibility or compliance
US5293870A (en) * 1989-11-17 1994-03-15 Board Of Regents The University Of Texas System Method and apparatus for elastographic measurement and imaging
US5474070A (en) * 1989-11-17 1995-12-12 The Board Of Regents Of The University Of Texas System Method and apparatus for elastographic measurement and imaging
US5353354A (en) * 1990-11-22 1994-10-04 Advanced Technology Laboratories, Inc. Acquisition and display of ultrasonic images from sequentially oriented image planes
US5257626A (en) * 1990-11-26 1993-11-02 Board Of Trustees Of The Leland Stanford Junior University Method of noninvasive myocardial motion analysis using bidirectional motion intergration in phase contrast MRI maps of myocardial velocity
US5262856A (en) * 1992-06-04 1993-11-16 Massachusetts Institute Of Technology Video image compositing techniques
US5305104A (en) * 1992-07-27 1994-04-19 The Trustees Of Columbia University In The City Of New York Digitally assisted motion compensated deinterlacing for enhanced definition television

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
An Electronic Video Camera Image Stabilizer Operated On Fuzzy Theory, by Yo Egusa et al., 1991 IEEE (0 7803 0236 Feb. 1992), pp. 851 858. *
An Electronic Video Camera Image Stabilizer Operated On Fuzzy Theory, by Yo Egusa et al., 1991 IEEE (0-7803-0236-Feb. 1992), pp. 851-858.
Kenya Uomori et al., "Automatic Image Stabilizing System By Full-Digital Signal Processing", IEEE Transactions on Consumer Electronics, vol. 36, No. 3, pp. 510-519, Aug. 1990.
Kenya Uomori et al., Automatic Image Stabilizing System By Full Digital Signal Processing , IEEE Transactions on Consumer Electronics , vol. 36, No. 3, pp. 510 519, Aug. 1990. *
Yo Egusa et al., "An Electronic Video Camera Image Stabilizer Operated on Fuzzy Theory", IEEE (journal unknown), pp. 851-858, 1992.
Yo Egusa et al., An Electronic Video Camera Image Stabilizer Operated on Fuzzy Theory , IEEE (journal unknown) , pp. 851 858, 1992. *

Cited By (183)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5899861A (en) * 1995-03-31 1999-05-04 Siemens Medical Systems, Inc. 3-dimensional volume by aggregating ultrasound fields of view
US5782766A (en) * 1995-03-31 1998-07-21 Siemens Medical Systems, Inc. Method and apparatus for generating and displaying panoramic ultrasound images
US6201900B1 (en) 1996-02-29 2001-03-13 Acuson Corporation Multiple ultrasound image registration system, method and transducer
US6014473A (en) 1996-02-29 2000-01-11 Acuson Corporation Multiple ultrasound image registration system, method and transducer
US6102865A (en) * 1996-02-29 2000-08-15 Acuson Corporation Multiple ultrasound image registration system, method and transducer
US6360027B1 (en) 1996-02-29 2002-03-19 Acuson Corporation Multiple ultrasound image registration system, method and transducer
US6222948B1 (en) 1996-02-29 2001-04-24 Acuson Corporation Multiple ultrasound image registration system, method and transducer
US6132376A (en) * 1996-02-29 2000-10-17 Acuson Corporation Multiple ultrasonic image registration system, method and transducer
US5956418A (en) * 1996-12-10 1999-09-21 Medsim Ltd. Method of mosaicing ultrasonic volumes for visual simulation
US6171248B1 (en) 1997-02-27 2001-01-09 Acuson Corporation Ultrasonic probe, system and method for two-dimensional imaging or three-dimensional reconstruction
US5876345A (en) * 1997-02-27 1999-03-02 Acuson Corporation Ultrasonic catheter, system and method for two dimensional imaging or three-dimensional reconstruction
US6045508A (en) * 1997-02-27 2000-04-04 Acuson Corporation Ultrasonic probe, system and method for two-dimensional imaging or three-dimensional reconstruction
US6152878A (en) * 1997-06-19 2000-11-28 Medinol Ltd. Intravascular ultrasound enhanced image and signal processing
US5876342A (en) * 1997-06-30 1999-03-02 Siemens Medical Systems, Inc. System and method for 3-D ultrasound imaging and motion estimation
US6116244A (en) * 1998-06-02 2000-09-12 Acuson Corporation Ultrasonic system and method for three-dimensional imaging with opacity control
US6755787B2 (en) 1998-06-02 2004-06-29 Acuson Corporation Medical diagnostic ultrasound system and method for versatile processing
US6685643B1 (en) * 1998-08-21 2004-02-03 Tomtec Imaging Systems Gmbh Method and device for recording ultrasonic images
US6224552B1 (en) 1998-10-01 2001-05-01 Atl Ultrasound Ultrasonic diagnostic imaging system with reduced spatial compounding seam artifacts
WO2000020887A1 (en) * 1998-10-01 2000-04-13 Koninklijke Philips Electronics N.V. Ultrasonic diagnostic imaging system with adaptive spatial compounding
US6126598A (en) * 1998-10-01 2000-10-03 Atl Ultrasound, Inc. Ultrasonic diagnostic imaging system with adaptive spatial compounding
US6547732B2 (en) 1998-10-01 2003-04-15 Koninklijke Philips Electronics N.V. Adaptive image processing for spatial compounding
US6117081A (en) * 1998-10-01 2000-09-12 Atl Ultrasound, Inc. Method for correcting blurring of spatially compounded ultrasonic diagnostic images
US6283917B1 (en) 1998-10-01 2001-09-04 Atl Ultrasound Ultrasonic diagnostic imaging system with blurring corrected spatial compounding
US6544177B1 (en) 1998-10-01 2003-04-08 Atl Ultrasound, Inc. Ultrasonic diagnostic imaging system and method with harmonic spatial compounding
US6159152A (en) * 1998-10-26 2000-12-12 Acuson Corporation Medical diagnostic ultrasound system and method for multiple image registration
US7837624B1 (en) * 1998-11-20 2010-11-23 Siemens Medical Solutions Usa, Inc. Medical diagnostic ultrasound imaging methods for extended field of view
US6554770B1 (en) 1998-11-20 2003-04-29 Acuson Corporation Medical diagnostic ultrasound imaging methods for extended field of view
US6641536B2 (en) 1998-11-20 2003-11-04 Acuson Corporation Medical diagnostic ultrasound imaging methods for extended field of view
US6364835B1 (en) 1998-11-20 2002-04-02 Acuson Corporation Medical diagnostic ultrasound imaging methods for extended field of view
US6101238A (en) * 1998-11-25 2000-08-08 Siemens Corporate Research, Inc. System for generating a compound x-ray image for diagnosis
US7756304B2 (en) 1998-12-30 2010-07-13 Siemens Medical Solutions Usa, Inc. Medical diagnostic ultrasonic imaging method and system for displaying multi-phase, multi-frame images
US20010012385A1 (en) * 1998-12-30 2001-08-09 John A Hossack Medical diagnostic ultrasonic imaging method and system for displaying multi-phase, multi-frame images
US6193657B1 (en) * 1998-12-31 2001-02-27 Ge Medical Systems Global Technology Company, Llc Image based probe position and orientation detection
EP1956555A1 (en) 1999-01-28 2008-08-13 Hitachi Ltd. Method and device of synthesizing panorama image from ultrasonic images
US6328693B1 (en) 1999-01-28 2001-12-11 Hitachi, Ltd. Method and device of synthesizing panorama image from ultrasound images
US6234968B1 (en) 1999-06-15 2001-05-22 Acuson Corporation 3-D diagnostic medical ultrasound imaging using a 1-D array
US6238345B1 (en) 1999-06-30 2001-05-29 Atl Ultrasound Image memory for extended field of view ultrasonic diagnostic imaging
US6730031B2 (en) 1999-06-30 2004-05-04 Koninklijke Philips Electronics, N.V. Editing and trimming apparatus and method for extended field of view ultrasonic imaging systems
US6299579B1 (en) 1999-06-30 2001-10-09 Atl Ultrasound Extended field of view ultrasonic diagnostic imaging with image reacquisition
US6442289B1 (en) 1999-06-30 2002-08-27 Koninklijke Philips Electronics N.V. Extended field of view ultrasonic diagnostic imaging
US6996177B1 (en) 1999-08-02 2006-02-07 Koninklijke Philips Electronics N.V. Motion estimation
US6254539B1 (en) 1999-08-26 2001-07-03 Acuson Corporation Transducer motion compensation in medical diagnostic ultrasound 3-D imaging
US6567469B1 (en) 2000-03-23 2003-05-20 Koninklijke Philips Electronics N.V. Motion estimation algorithm suitable for H.261 videoconferencing applications
US6390981B1 (en) 2000-05-23 2002-05-21 Koninklijke Philips Electronics N.V. Ultrasonic spatial compounding with curved array scanheads
US6423004B1 (en) * 2000-05-30 2002-07-23 Ge Medical Systems Global Technology Company, Llc Real-time ultrasound spatial compounding using multiple angles of view
US6416477B1 (en) 2000-08-22 2002-07-09 Koninklijke Philips Electronics N.V. Ultrasonic diagnostic systems with spatial compounded panoramic imaging
US6872181B2 (en) 2001-04-25 2005-03-29 Siemens Medical Solutions Usa, Inc. Compound image display system and method
US20020167533A1 (en) * 2001-04-25 2002-11-14 Siemens Medical Systems, Inc. Compound image display system and method
US6605042B2 (en) 2001-08-10 2003-08-12 Ge Medical Systems Global Technology Company, Llc Method and apparatus for rotation registration of extended field of view ultrasound images
US6503201B1 (en) 2001-10-03 2003-01-07 Koninklijke Philips Electronics N.V. Correction of extended field of view images for distortion due to scanhead motion
US6589176B2 (en) 2001-12-05 2003-07-08 Koninklijke Philips Electronics N.V. Ultrasonic image stabilization system and method
US6572549B1 (en) 2001-12-18 2003-06-03 Koninklijke Philips Electronics Nv High frame rate extended field of view ultrasound imaging system and method
US20040127796A1 (en) * 2002-06-07 2004-07-01 Vikram Chalana 3D ultrasound-based instrument for non-invasive measurement of amniotic fluid volume
US7744534B2 (en) 2002-06-07 2010-06-29 Verathon Inc. 3D ultrasound-based instrument for non-invasive measurement of amniotic fluid volume
US20080146932A1 (en) * 2002-06-07 2008-06-19 Vikram Chalana 3D ultrasound-based instrument for non-invasive measurement of Amniotic Fluid Volume
US20060235301A1 (en) * 2002-06-07 2006-10-19 Vikram Chalana 3D ultrasound-based instrument for non-invasive measurement of amniotic fluid volume
US7819806B2 (en) 2002-06-07 2010-10-26 Verathon Inc. System and method to identify and measure organ wall boundaries
US8221321B2 (en) 2002-06-07 2012-07-17 Verathon Inc. Systems and methods for quantification and classification of fluids in human cavities in ultrasound images
US8221322B2 (en) 2002-06-07 2012-07-17 Verathon Inc. Systems and methods to improve clarity in ultrasound images
US7087022B2 (en) 2002-06-07 2006-08-08 Diagnostic Ultrasound Corporation 3D ultrasound-based instrument for non-invasive measurement of amniotic fluid volume
US8308644B2 (en) 2002-08-09 2012-11-13 Verathon Inc. Instantaneous ultrasonic measurement of bladder volume
US9993225B2 (en) 2002-08-09 2018-06-12 Verathon Inc. Instantaneous ultrasonic echo measurement of bladder volume with a limited number of ultrasound beams
US8167802B2 (en) * 2002-09-12 2012-05-01 Hitachi Medical Corporation Biological tissue motion trace method and image diagnosis device using the trace method
US20060173292A1 (en) * 2002-09-12 2006-08-03 Hirotaka Baba Biological tissue motion trace method and image diagnosis device using the trace method
US7356172B2 (en) * 2002-09-26 2008-04-08 Siemens Medical Solutions Usa, Inc. Methods and systems for motion tracking
US20040064036A1 (en) * 2002-09-26 2004-04-01 Zuhua Mao Methods and systems for motion tracking
US6867753B2 (en) * 2002-10-28 2005-03-15 University Of Washington Virtual image registration in augmented display field
US20040080467A1 (en) * 2002-10-28 2004-04-29 University Of Washington Virtual image registration in augmented display field
WO2004063990A1 (en) * 2003-01-13 2004-07-29 Koninklijke Philips Electronics N.V. A method of image registration and medical image data processing apparatus
JP4744883B2 (en) * 2003-01-13 2011-08-10 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Image alignment method and medical image data processing apparatus
EP3239930A1 (en) * 2003-01-13 2017-11-01 Koninklijke Philips N.V. A method of image registration and medical image data processing apparatus
US20060188134A1 (en) * 2003-01-13 2006-08-24 Quist Marcel J Method of image registration and medical image data processing apparatus
US7778490B2 (en) 2003-01-13 2010-08-17 Koninklijke Philips Electronics N.V. Method of image registration and medical image data processing apparatus
JP2006512960A (en) * 2003-01-13 2006-04-20 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Image alignment method and medical image data processing apparatus
US6858010B2 (en) 2003-05-06 2005-02-22 Siemens Medical Solutions Usa, Inc. Identifying clinical markers in spatial compounding ultrasound imaging
US7033320B2 (en) 2003-08-05 2006-04-25 Siemens Medical Solutions Usa, Inc. Extended volume ultrasound data acquisition
US20050033173A1 (en) * 2003-08-05 2005-02-10 Von Behren Patrick L. Extended volume ultrasound data acquisition
US20050096539A1 (en) * 2003-10-31 2005-05-05 Siemens Medical Solutions Usa, Inc. Intelligent ultrasound examination storage system
US7658714B2 (en) 2003-10-31 2010-02-09 Siemens Medical Solutions Usa, Inc. Intelligent ultrasound examination storage system
US20050259737A1 (en) * 2004-05-18 2005-11-24 Sunplus Technology Co., Ltd. Optimal correlation matching method and system for determining track behavior
US7903738B2 (en) * 2004-05-18 2011-03-08 Sunplus Technology Co., Ltd. Optimal correlation matching method and system for determining track behavior
US20060013501A1 (en) * 2004-07-09 2006-01-19 Tadashi Tamura Method and apparatus for image processing
US7386183B2 (en) * 2004-07-09 2008-06-10 Aloka Co., Ltd. Method and apparatus for image processing
US7684628B2 (en) * 2004-08-12 2010-03-23 Industrial Technology Research Institute System and method for non-iterative global motion estimation
US20060034528A1 (en) * 2004-08-12 2006-02-16 Yeping Su System and method for non-iterative global motion estimation
US8711366B2 (en) 2005-01-21 2014-04-29 Carl Zeiss Meditec, Inc. Method of motion correction in optical coherence tomography imaging
US9167964B2 (en) 2005-01-21 2015-10-27 Carl Zeiss Meditec, Inc. Method of motion correction in optical coherence tomography imaging
US9706915B2 (en) 2005-01-21 2017-07-18 Carl Zeiss Meditec, Inc. Method of motion correction in optical coherence tomography imaging
CN100455266C (en) * 2005-03-29 2009-01-28 深圳迈瑞生物医疗电子股份有限公司 Broad image processing method
US8050504B2 (en) 2005-04-06 2011-11-01 Carl Zeis Meditec, Inc. Method and apparatus for measuring motion of a subject using a series of partial images from an imaging system
US8363958B2 (en) 2005-04-06 2013-01-29 Carl Zeiss Meditec, Inc. Method and apparatus for measuring motion of a subject using a series of partial images from an imaging system
US8649611B2 (en) 2005-04-06 2014-02-11 Carl Zeiss Meditec, Inc. Method and apparatus for measuring motion of a subject using a series of partial images from an imaging system
US20110043757A1 (en) * 2005-04-06 2011-02-24 Carl Zeiss Meditec, Inc. Method and apparatus for measuring motion of a subject using a series of partial images from an imaging system
US20060228011A1 (en) * 2005-04-06 2006-10-12 Everett Matthew J Method and apparatus for measuring motion of a subject using a series of partial images from an imaging system
US9033504B2 (en) 2005-04-06 2015-05-19 Carl Zeiss Meditec, Inc. Method and apparatus for measuring motion of a subject using a series of partial images from an imaging system
US7805009B2 (en) 2005-04-06 2010-09-28 Carl Zeiss Meditec, Inc. Method and apparatus for measuring motion of a subject using a series of partial images from an imaging system
US9198600B2 (en) 2005-05-06 2015-12-01 Vasonova, Inc. Endovascular access and guidance system utilizing divergent beam ultrasound
US9339207B2 (en) 2005-05-06 2016-05-17 Vasonova, Inc. Endovascular devices and methods of use
US10321890B2 (en) 2005-05-06 2019-06-18 Arrow International, Inc. Apparatus and method for endovascular device guiding and positioning using physiological parameters
US10368837B2 (en) 2005-05-06 2019-08-06 Arrow International, Inc. Apparatus and method for vascular access
US10470743B2 (en) 2005-05-06 2019-11-12 Arrow International, Inc. Apparatus and method for endovascular device guiding and positioning using physiological parameters
US8409103B2 (en) 2005-05-06 2013-04-02 Vasonova, Inc. Ultrasound methods of positioning guided vascular access devices in the venous system
US8597193B2 (en) 2005-05-06 2013-12-03 Vasonova, Inc. Apparatus and method for endovascular device guiding and positioning using physiological parameters
US9204819B2 (en) 2005-05-06 2015-12-08 Vasonova, Inc. Endovenous access and guidance system utilizing non-image based ultrasound
US10335240B2 (en) 2005-05-06 2019-07-02 Arrow International, Inc. Endovascular navigation system and method
US8275185B2 (en) 2005-11-10 2012-09-25 Microsoft Corporation Discover biological features using composite images
US20070211928A1 (en) * 2005-11-10 2007-09-13 Rosetta Inpharmatics Llc Discover biological features using composite images
US20110110569A1 (en) * 2005-11-10 2011-05-12 Microsoft Corporation Discover biological features using composite images
US7894650B2 (en) 2005-11-10 2011-02-22 Microsoft Corporation Discover biological features using composite images
US20070255137A1 (en) * 2006-05-01 2007-11-01 Siemens Medical Solutions Usa, Inc. Extended volume ultrasound data display and measurement
US20080021317A1 (en) * 2006-07-24 2008-01-24 Siemens Medical Solutions Usa, Inc. Ultrasound medical imaging with robotic assistance for volume imaging
US8167803B2 (en) 2007-05-16 2012-05-01 Verathon Inc. System and method for bladder detection using harmonic imaging
US8133181B2 (en) 2007-05-16 2012-03-13 Verathon Inc. Device, system and method to measure abdominal aortic aneurysm diameter
US7780601B2 (en) 2007-06-05 2010-08-24 Siemens Medical Solutions Usa, Inc. Adaptive clinical marker preservation in spatial compound ultrasound imaging
US20100160783A1 (en) * 2008-12-19 2010-06-24 Menachem Halmann Ultrasound imaging method and apparatus
US9445780B2 (en) 2009-12-04 2016-09-20 University Of Virginia Patent Foundation Tracked ultrasound vessel imaging
US11445996B2 (en) 2010-11-08 2022-09-20 Teleflex Life Sciences Limited Endovascular navigation system and method
US9119551B2 (en) 2010-11-08 2015-09-01 Vasonova, Inc. Endovascular navigation system and method
US10368830B2 (en) 2010-11-08 2019-08-06 Arrow International Inc. Endovascular navigation system and method
US9164171B2 (en) * 2011-02-21 2015-10-20 Samsung Electronics Co., Ltd. Method of ultrasonic beamforming and apparatus therefor
US20120212618A1 (en) * 2011-02-21 2012-08-23 Samsung Electronics Co., Ltd. Method of ultrasonic beamforming and apparatus therefor
US10092178B2 (en) 2011-03-30 2018-10-09 Carl Zeiss Meditec, Inc. Systems and methods for efficiently obtaining measurements of the human eye using tracking
US9033510B2 (en) 2011-03-30 2015-05-19 Carl Zeiss Meditec, Inc. Systems and methods for efficiently obtaining measurements of the human eye using tracking
US8857988B2 (en) 2011-07-07 2014-10-14 Carl Zeiss Meditec, Inc. Data acquisition methods for reduced motion artifacts and applications in OCT angiography
US8897521B2 (en) * 2011-08-19 2014-11-25 Industrial Technology Research Institute Ultrasound image registration apparatus and method thereof
US20130044929A1 (en) * 2011-08-19 2013-02-21 Industrial Technology Research Institute Ultrasound image registration apparatus and method thereof
US9706914B2 (en) 2012-01-19 2017-07-18 Carl Zeiss Meditec, Inc. Systems and methods for enhanced accuracy in OCT imaging of the cornea
US8891840B2 (en) 2012-02-13 2014-11-18 Siemens Medical Solutions Usa, Inc. Dynamic steered spatial compounding in ultrasound imaging
US9081097B2 (en) 2012-05-01 2015-07-14 Siemens Medical Solutions Usa, Inc. Component frame enhancement for spatial compounding in ultrasound imaging
US9743994B2 (en) 2012-05-07 2017-08-29 Vasonova, Inc. Right atrium indicator
US9345447B2 (en) 2012-05-07 2016-05-24 Vasonova, Inc. Right atrium indicator
US8965490B2 (en) 2012-05-07 2015-02-24 Vasonova, Inc. Systems and methods for detection of the superior vena cava area
US20150342560A1 (en) * 2013-01-25 2015-12-03 Ultrasafe Ultrasound Llc Novel Algorithms for Feature Detection and Hiding from Ultrasound Images
US11298109B2 (en) * 2013-03-11 2022-04-12 Canon Medical Systems Corporation Ultrasonic diagnostic apparatus and image processing apparatus
US20150063628A1 (en) * 2013-09-04 2015-03-05 Xerox Corporation Robust and computationally efficient video-based object tracking in regularized motion environments
US9213901B2 (en) * 2013-09-04 2015-12-15 Xerox Corporation Robust and computationally efficient video-based object tracking in regularized motion environments
US10670716B2 (en) 2016-05-04 2020-06-02 Invensense, Inc. Operating a two-dimensional array of ultrasonic transducers
US10656255B2 (en) 2016-05-04 2020-05-19 Invensense, Inc. Piezoelectric micromachined ultrasonic transducer (PMUT)
US10315222B2 (en) 2016-05-04 2019-06-11 Invensense, Inc. Two-dimensional array of CMOS control elements
US10445547B2 (en) 2016-05-04 2019-10-15 Invensense, Inc. Device mountable packaging of ultrasonic transducers
US11440052B2 (en) 2016-05-04 2022-09-13 Invensense, Inc. Two-dimensional array of CMOS control elements
US11651611B2 (en) 2016-05-04 2023-05-16 Invensense, Inc. Device mountable packaging of ultrasonic transducers
US10325915B2 (en) 2016-05-04 2019-06-18 Invensense, Inc. Two-dimensional array of CMOS control elements
US10706835B2 (en) 2016-05-10 2020-07-07 Invensense, Inc. Transmit beamforming of a two-dimensional array of ultrasonic transducers
US10441975B2 (en) 2016-05-10 2019-10-15 Invensense, Inc. Supplemental sensor modes and systems for ultrasonic transducers
US11288891B2 (en) 2016-05-10 2022-03-29 Invensense, Inc. Operating a fingerprint sensor comprised of ultrasonic transducers
US10600403B2 (en) 2016-05-10 2020-03-24 Invensense, Inc. Transmit operation of an ultrasonic sensor
US10632500B2 (en) 2016-05-10 2020-04-28 Invensense, Inc. Ultrasonic transducer with a non-uniform membrane
US10408797B2 (en) 2016-05-10 2019-09-10 Invensense, Inc. Sensing device with a temperature sensor
US10539539B2 (en) 2016-05-10 2020-01-21 Invensense, Inc. Operation of an ultrasonic sensor
US11471912B2 (en) 2016-05-10 2022-10-18 Invensense, Inc. Supplemental sensor modes and systems for ultrasonic transducers
US11626099B2 (en) 2016-05-10 2023-04-11 Invensense, Inc. Transmit beamforming of a two-dimensional array of ultrasonic transducers
US11154906B2 (en) 2016-05-10 2021-10-26 Invensense, Inc. Receive operation of an ultrasonic sensor
US11673165B2 (en) 2016-05-10 2023-06-13 Invensense, Inc. Ultrasonic transducer operable in a surface acoustic wave (SAW) mode
US10452887B2 (en) 2016-05-10 2019-10-22 Invensense, Inc. Operating a fingerprint sensor comprised of ultrasonic transducers
US11112388B2 (en) 2016-05-10 2021-09-07 Invensense, Inc. Operation of an ultrasonic sensor
US10562070B2 (en) 2016-05-10 2020-02-18 Invensense, Inc. Receive operation of an ultrasonic sensor
US11185305B2 (en) * 2016-06-30 2021-11-30 Koninklijke Philips N.V. Intertial device tracking system and method of operation thereof
WO2018046455A1 (en) 2016-09-09 2018-03-15 Koninklijke Philips N.V. Stabilization of ultrasound images
US11712225B2 (en) 2016-09-09 2023-08-01 Koninklijke Philips N.V. Stabilization of ultrasound images
CN109937370A (en) * 2016-09-09 2019-06-25 皇家飞利浦有限公司 The stabilization of ultrasound image
US10891461B2 (en) 2017-05-22 2021-01-12 Invensense, Inc. Live fingerprint detection utilizing an integrated ultrasound and infrared sensor
US10474862B2 (en) * 2017-06-01 2019-11-12 Invensense, Inc. Image generation in an electronic device using ultrasonic transducers
US10860831B2 (en) * 2017-06-01 2020-12-08 Invensense, Inc. Image generation in an electronic device using ultrasonic transducers
US20200074135A1 (en) * 2017-06-01 2020-03-05 Invensense, Inc. Image generation in an electronic device using ultrasonic transducers
US10643052B2 (en) 2017-06-28 2020-05-05 Invensense, Inc. Image generation in an electronic device using ultrasonic transducers
US10997388B2 (en) 2017-12-01 2021-05-04 Invensense, Inc. Darkfield contamination detection
US10936841B2 (en) 2017-12-01 2021-03-02 Invensense, Inc. Darkfield tracking
US10984209B2 (en) 2017-12-01 2021-04-20 Invensense, Inc. Darkfield modeling
US11151355B2 (en) 2018-01-24 2021-10-19 Invensense, Inc. Generation of an estimated fingerprint
US10755067B2 (en) 2018-03-22 2020-08-25 Invensense, Inc. Operating a fingerprint sensor comprised of ultrasonic transducers
US10936843B2 (en) 2018-12-28 2021-03-02 Invensense, Inc. Segmented image acquisition
US11188735B2 (en) 2019-06-24 2021-11-30 Invensense, Inc. Fake finger detection using ridge features
US11216681B2 (en) 2019-06-25 2022-01-04 Invensense, Inc. Fake finger detection based on transient features
US11216632B2 (en) 2019-07-17 2022-01-04 Invensense, Inc. Ultrasonic fingerprint sensor with a contact layer of non-uniform thickness
US11176345B2 (en) 2019-07-17 2021-11-16 Invensense, Inc. Ultrasonic fingerprint sensor with a contact layer of non-uniform thickness
US11682228B2 (en) 2019-07-17 2023-06-20 Invensense, Inc. Ultrasonic fingerprint sensor with a contact layer of non-uniform thickness
US11232549B2 (en) 2019-08-23 2022-01-25 Invensense, Inc. Adapting a quality threshold for a fingerprint image
US11392789B2 (en) 2019-10-21 2022-07-19 Invensense, Inc. Fingerprint authentication using a synthetic enrollment image
US11460957B2 (en) 2020-03-09 2022-10-04 Invensense, Inc. Ultrasonic fingerprint sensor with a contact layer of non-uniform thickness
US11243300B2 (en) 2020-03-10 2022-02-08 Invensense, Inc. Operating a fingerprint sensor comprised of ultrasonic transducers and a presence sensor
US11328165B2 (en) 2020-04-24 2022-05-10 Invensense, Inc. Pressure-based activation of fingerprint spoof detection
CN112012062A (en) * 2020-09-01 2020-12-01 天津动网信息科技有限公司 Rail and roadbed offset three-coordinate monitoring device and method

Also Published As

Publication number Publication date
DE19611990C2 (en) 2003-08-28
DE19611990A1 (en) 1996-10-02
JPH08280688A (en) 1996-10-29
JP3345257B2 (en) 2002-11-18

Similar Documents

Publication Publication Date Title
US5575286A (en) Method and apparatus for generating large compound ultrasound image
US5566674A (en) Method and apparatus for reducing ultrasound image shadowing and speckle
US5782766A (en) Method and apparatus for generating and displaying panoramic ultrasound images
US5655535A (en) 3-Dimensional compound ultrasound field of view
US5899861A (en) 3-dimensional volume by aggregating ultrasound fields of view
US5876342A (en) System and method for 3-D ultrasound imaging and motion estimation
EP0883860B1 (en) Multiple ultrasound image registration system, method and transducer
US6475149B1 (en) Border detection method and system
US20190216435A1 (en) Sensor coordinate calibration in an ultrasound system
EP0880937B1 (en) Ultrasonic diagnostic imaging system with doppler assisted tracking of tissue motion
US6442289B1 (en) Extended field of view ultrasonic diagnostic imaging
US6872181B2 (en) Compound image display system and method
EP1024455B1 (en) Method and device of synthesizing panorama image from ultrasonic images
Herlin et al. Features extraction and analysis methods for sequences of ultrasound images
US6659953B1 (en) Morphing diagnostic ultrasound images for perfusion assessment
KR100961856B1 (en) Ultrasound system and method for forming ultrasound image
US6517488B1 (en) Medical diagnostic ultrasound system and method for identifying constrictions
US6238345B1 (en) Image memory for extended field of view ultrasonic diagnostic imaging
US6540681B1 (en) Extended view ultrasound imaging system
EP2345992A2 (en) Ultrasound image processing based on motion degree of ultrasound probe
US5623929A (en) Ultrasonic doppler flow imaging method for eliminating motion artifacts
US10278663B2 (en) Sensor coordinate calibration in an ultrasound system
US20240050062A1 (en) Analyzing apparatus and analyzing method
Sonka et al. Three-dimensional automated segmentation of coronary wall and plaque from intravascular ultrasound pullback sequences
EP2168494A1 (en) Ultrasound volume data processing

Legal Events

Date Code Title Description
AS Assignment

Owner name: SIEMENS MEDICAL SYSTEMS, INC., NEW JERSEY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WENG, LEE;TIRUMALAI, CHUN Q.;REEL/FRAME:007424/0260

Effective date: 19950330

STCF Information on status: patent grant

Free format text: PATENTED CASE

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FPAY Fee payment

Year of fee payment: 4

FPAY Fee payment

Year of fee payment: 8

FPAY Fee payment

Year of fee payment: 12

AS Assignment

Owner name: SIEMENS MEDICAL SOLUTIONS USA, INC.,PENNSYLVANIA

Free format text: CHANGE OF NAME;ASSIGNOR:SIEMENS MEDICAL SYSTEMS, INC.;REEL/FRAME:024529/0474

Effective date: 20010801