US20050213845A1 - Method and product for processing digital images - Google Patents

Method and product for processing digital images Download PDF

Info

Publication number
US20050213845A1
US20050213845A1 US10/708,771 US70877104A US2005213845A1 US 20050213845 A1 US20050213845 A1 US 20050213845A1 US 70877104 A US70877104 A US 70877104A US 2005213845 A1 US2005213845 A1 US 2005213845A1
Authority
US
United States
Prior art keywords
region
foreground
transition
background
function
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US10/708,771
Other versions
US7623728B2 (en
Inventor
Gopal Avinash
Rakesh Lal
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
General Electric Co
Original Assignee
General Electric Co
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by General Electric Co filed Critical General Electric Co
Priority to US10/708,771 priority Critical patent/US7623728B2/en
Assigned to GENERAL ELECTRIC COMPANY reassignment GENERAL ELECTRIC COMPANY ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: AVINASH, GOPAL B., LAL, RAKESH MOHAN
Publication of US20050213845A1 publication Critical patent/US20050213845A1/en
Application granted granted Critical
Publication of US7623728B2 publication Critical patent/US7623728B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • G06T5/70
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/194Segmentation; Edge detection involving foreground-background segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10088Magnetic resonance imaging [MRI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20004Adaptive image processing
    • G06T2207/20012Locally adaptive
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing

Definitions

  • the present disclosure relates generally to a method and product for processing digital images, and particularly to a method and product for processing digital images with detection and suppression of background noise.
  • a digital image is a collection of digital information in the form of pixels that may be processed to provide a visual image representative of an object-of-interest.
  • the digital image may be obtained from digital cameras, digital video, digital scanners, or the like, or may be digitized from an analog film.
  • An exemplary object-of-interest is a biological object
  • an exemplary digital scanner is a magnetic resonance imaging (MRI) scanner. While reference may be made herein to MRI as an exemplary scanning method, it will be appreciated that other scanning methods may be employed, such as computed tomography (CT), ultrasound, or X-ray, for example.
  • CT computed tomography
  • ultrasound ultrasound
  • X-ray X-ray
  • noise may be prominent throughout an image, it is most prominent in the background regions where the signal is expected to be zero.
  • the presence of background intensity in an image unnecessarily distracts an observer who may be viewing the final image. Accordingly, there is a need in the art for a digital image processing method and product that improves image quality by detecting and suppressing background noise.
  • Embodiments of the invention include a method for processing a digital image.
  • a foreground region relating to an imaged object is estimated, a background region relating to other than the imaged object is estimated, and by using the image, the estimated foreground region and the estimated background region, a transition region disposed between the foreground region and the background region is calculated.
  • the estimated foreground region, the estimated background region, and the calculated transition region each include a separate set of pixels that may each be processed separately for suppressing pixel intensities in the estimated background region and improving image quality.
  • inventions include a computer program product for processing a digital image, the product including a storage medium, readable by a processing circuit, storing instructions for execution by the processing circuit for performing embodiments of the aforementioned method.
  • FIG. 1 depicts an exemplary method and system for processing digital images in accordance with embodiments of the invention
  • FIG. 2 depicts further detail of the method of FIG. 1 ;
  • FIG. 3 depicts further detail of the method of FIG. 2 ;
  • FIG. 4 depicts an alternative method to that depicted in FIG. 3 ;
  • FIG. 5 depicts yet further detail of the method of FIG. 1 .
  • Embodiments of the invention provide a method and product for processing digital images, and particularly to a method and product for processing digital images with detection and suppression of background noise.
  • a biological object is scanned using an MRI imaging technique, and the resulting image is digitally processed for the detection of background noise and suppression thereof, thereby improving the signal-to-noise-ratio (SNR) of the resulting image.
  • the digital image is separated into foreground, background and transition regions, thereby enabling each region to be analyzed and processed separately.
  • the processing techniques applied involve low-level image processing techniques, such as thresholding and binary image subtraction for example. While embodiments described herein may employ MRI as an exemplary imaging technique, it will be appreciated that the disclosed invention may also employ other imaging techniques, such as CT, Ultrasound, X-ray, and the like, for example.
  • FIG. 1 a high-level flowchart of an exemplary method 100 for processing a digital image is depicted.
  • An original image 110 is analyzed and segmented at block 120 into foreground, background and transition regions, with the background region being referred to as a background mask 130 , and the foreground and transition regions being incorporated into and referred to as a transition mask 140 .
  • a background detection method 200 discussed hereinafter in reference to FIG. 2 , segments original image 110 into the three aforementioned regions. While reference is made herein to an “original” image 110 , it will be appreciated that this reference is merely a reference to the image entering analysis block 120 , and does not infer a “first occurrence” of image 110 .
  • original image 110 may be a digital representation of an analog film resulting from an imaging technique, such as X-ray for example.
  • the analog film would be considered the “first occurrence” image
  • the digital representation would be considered the original image 110 entering analysis block 120 .
  • a background noise suppression method 300 discussed hereinafter in reference to FIG. 3 , for suppressing the background noise and improving the SNR of the resulting image.
  • the resulting image is then filtered to produce a filtered image 150 . While embodiments described herein disclose the suppression of background noise, it will be appreciated that such suppression may generally be described as the suppression of pixel intensities in the background region.
  • a foreground region relating to the imaged biological object is estimated 202 by applying a pixel intensity threshold t_high 402 to filter out all but those pixels having an intensity equal to or greater than t_high 402 , where such filtering may alternatively be viewed as filtering original image 110 according to a first criterion.
  • the value of t_high 402 is set to be about 50% of the mean intensity of the non-zero pixels in original image 110 .
  • the resulting region is referred to as the estimated foreground region 204 , also referred to as the initial foreground region.
  • a background region relating to other than the imaged object is estimated as containing all pixels in original image 110 having an intensity of zero.
  • the resulting region is referred to as the estimated background region 206 , which is also referred to as the zero mask.
  • the estimated background region 206 may be defined as that region containing those pixels of the image meeting a second criterion, such as an intensity of zero, or an intensity less than a low intensity threshold, for example.
  • a gradient magnitude image 208 is computed that provides a value for the gradient of the intensity of each pixel of original image 110 .
  • a gradient-constrained-hysteresis-threshold (GCHT) method 400 is applied to original image 110 using gradient magnitude image 208 to calculate an initial transition region 210 .
  • Initial transition region 210 is disposed between estimated foreground and background regions 204 , 206 , which may be seen by referring to the illustration depicted FIG. 5 .
  • initial transition region 210 is calculated to be that region containing pixels having a morphological connection (that is, a connection made possible via a morphological operation) to a pixel of estimated foreground region 204 , having an intensity greater than a low threshold t_low 404 , and having a gradient magnitude that is within a gradient tolerance value g_tol 406 of the gradient magnitude of the foreground pixel to which it is connected, which is depicted in algorithm form in FIGS. 3 and 4 .
  • initial transition region 210 may be considered to be that region of the image containing those pixels that meet a transition region criterion, such as the set of aforementioned criteria, for example.
  • GCHT method 400 with hysteresis thresholding and gradient constraining is depicted in the absence of an iterative approach.
  • the inputs to GCHT method 400 are gradient magnitude image 208 and estimated foreground region 204 , where estimated foreground region 204 has been estimated using intensity threshold t_high 402 .
  • the output of GCHT method 400 is a gradient constrained hysteresis threshold image 450 . For each pixel x in the foreground (estimated foreground region 204 for example) where the pixel intensity Ix is greater than t_high 402 , initial transition region 210 is calculated to be that region defined by the set of pixels y where:
  • the values for low threshold t_low 404 and gradient tolerance value g_tol 406 may be user adjusted, thereby making GCHT method 400 tunable.
  • Estimated foreground region 204 , estimated background region 206 , and calculated initial transition region 210 each comprise a separate set of pixels that may each be processed separately for suppressing background noise and improving image quality.
  • the y pixels of initial transition region 210 may be added to the x pixels of estimated foreground region 204 , thereby growing the foreground region that is representative of the biological object being imaged for greater object renditioning.
  • initial transition region 210 becomes transition mask 140 .
  • GCHT method 400 An advantage associated with GCHT method 400 is that in an image having strong edges, the growing region grows along the edge and not across the edge. In this way, moderate intensity background artifacts that are connected to an edge will not get incorporated into the transition region. As a result, the t_low threshold 404 may be set quite low to capture low intensity foreground regions into the transition region without incorporating artifacts with higher intensities that are in the background. Also, this approach will capture the low intensity foreground regions in images that suffer from intensity variation as long as the image gradient in the regions of intensity variation is less than the gradient tolerance value g_tol.
  • GCHT method 400 is modified to include an iterative approach, thereby resulting in an iterative GCHT method 500 .
  • GCHT method 400 is shown as one operation in an iterative do-loop 505 .
  • the gradient tolerance value g_tol 406 is set 510 to a low gradient tolerance value tol_low 408 , after which, GCHT method 400 is applied as discussed previously.
  • GCHT method 400 is applied as discussed previously.
  • it is determined whether the foreground-plus-transition-region (for example, estimated foreground region 204 plus initial transition region 210 for the first pass in the iterative process) has grown by more than a defined number of pixels size_t. In response to the determination at 520 being yes, all pixels that may be added to the foreground since the last labeling are labeled as incremental transition region “i”, and the iteration counter is incremented by one (i i+1). These actions are depicted at 530 .
  • gradient tolerance value g_tol 406 is incremented by a gradient tolerance value step g_tol_step 410 until a high gradient tolerance value tol_high 412 condition is met or exceeded.
  • steps g_tol_step 410 are depicted at 540 and 550 , respectively, where 550 is labeled “repeat while: g_tol ⁇ tol_high”.
  • g_tol ⁇ tol_high is labeled “repeat while: g_tol ⁇ tol_high”.
  • incremental transitions regions i, i+1, i+2, etc.
  • each incremental transition region having an incrementally larger gradient tolerance value g_tol 406 that has a value between tol_low 408 and tol_high 412 .
  • the iterative do-loop 505 is continued until tol_high 412 is met or exceeded.
  • each incremental transition region (i, i+1, i+2, etc.) 530 is calculated to be that region containing pixels connected to a pixel of a previously calculated incremental transition region 530 , having an intensity greater than a low threshold t_low 404 , and having a gradient magnitude that is within the incrementally larger gradient tolerance value (g_tol+g_tol_step) of the gradient magnitude of the incremental transition region pixel to which it is connected.
  • the adjustment of single transition region 560 by focus parameter 570 results in transition mask 140 .
  • An advantage of using a focus parameter is that the internal parameters of method 100 may be fixed and the entire process may be made controllable or tunable by the single focus parameter 570 . The higher the focus parameter 570 , the more liberal method 100 is in accepting pixels into transition mask 140 .
  • the following parameters may be tunable:
  • low gradient tolerance value tol_low 408 and high gradient tolerance value tol_high 412 may be set to a specific percentage of the average gradient magnitude of the entire original image 110
  • gradient tolerance value step g_tol_step 410 may be set according to a desired number of iterations, such as (tol_high-tol_low)/10 where ten iterations are desired for generating ten incremental transition regions 530 , for example.
  • an object region also referred to as the object image
  • the term object region refers to those pixels of the original image that are determined via method 100 to be representative of the biological object under observation.
  • Such morphological operations may include erosion of the image and dilation back to the transition mask to remove small objects and bridges that may connect background noise to the main object, and dilation of the image and erosion back to the original size to fill small cracks in the edges of the image 212 .
  • the zero mask may then be combined with the object image, which may then undergo a connected-components morphological operation to fill small holes in the resulting image 214 .
  • the zero mask is removed from the image resulting in the final object mask 216 .
  • the final object mask 216 is then separated into a final foreground region and a final transition region 218 .
  • the initial estimated foreground region 204 is labeled the final foreground mask 152
  • the difference between the object region and the final foreground mask is labeled the final transition mask 154
  • the remainder of the image is labeled the final background mask 156 , which collectively make up the filtered image 150 , best seen by referring back to FIG. 1 .
  • the intensities of the foreground pixels may be increased slightly to improve contrast and increase SNR in the foreground regions, while the intensities of the background pixels are lowered to reduce noise in the background regions.
  • the final object image is the union of the final foreground mask 152 and the final transition mask 154 after filtering.
  • the background noise may be suppressed gradually by reducing the intensity of the background pixels to zero as a function of their distance from the final object image.
  • final foreground mask 152 final transition mask 154
  • final background mask 156 are depicted. Shown at increasing distances from final transition mask 154 are two regions 310 , 320 , which extend into the background region defined by final background mask 156 .
  • the background noise in regions 310 , 320 is suppressed gradually by reducing the intensity of the background pixels in those regions to zero the further the region is from final transition mask 154 .
  • Different functions may be employed to gradually reduce the intensity of the background pixels to zero, such as a Gaussian function, a Hanning function, a Hamming function, or any function suitable for reducing a pixel intensity value with respect to distance from a defined origin. Combinations of functions may be employed as desired. Alternatively, all background pixel intensities may be identically set to zero, or the background region may be homogenized by iteratively blurring the background region and adding a small fraction of the original image back into the background region.
  • Embodiments of the invention may be provided in executable instruction form on a storage medium, such as memory 605 or in the form of a hard drive, a floppy disk or a CD-ROM 610 for example, that is readable by a processing circuit, such as processor 615 for example, the processing circuit being in signal communication via application software with a graphical user interface at a computer, such as computer 600 for example, whereby a user may execute the embedded instructions for practicing the disclosed invention.
  • the instructions may be loaded into and/or executed by computer 600 , or transmitted over some transmission medium, such as over electrical wiring or cabling, through fiber optics, or via electromagnetic radiation, wherein when the computer program code is loaded into and executed by computer 600 , the computer becomes an apparatus for practicing the invention.
  • the computer program code segments When implemented on a general-purpose microprocessor, the computer program code segments configure the microprocessor to create specific logic circuits.
  • the technical effect of the executable instructions is to enhance the quality of a digital image, of a scanned biological object for example, by detecting and suppressing background noise in the image.
  • some embodiments of the invention may include some of the following advantages: use of low level image processing tools (no statistical models, no neural network based methods, for example) for improving image quality; use of method parameters that may be tunable or predefined; the ability to capture low intensity foreground regions into the transition mask without incorporating artifacts with higher intensities that are in the background; and, the ability to execute morphological and filtering operations on discrete regions of the image thereby providing greater control over the generation and quality of the final object image.
  • low level image processing tools no statistical models, no neural network based methods, for example
  • method parameters that may be tunable or predefined
  • the ability to capture low intensity foreground regions into the transition mask without incorporating artifacts with higher intensities that are in the background and, the ability to execute morphological and filtering operations on discrete regions of the image thereby providing greater control over the generation and quality of the final object image.

Abstract

A method and computer program product for processing a digital image is disclosed. A foreground region relating to an imaged object is estimated, a background region relating to other than the imaged object is estimated, and by using the image, the estimated foreground region and the estimated background region, a transition region disposed between the foreground region and the background region is calculated. The estimated foreground region, the estimated background region, and the calculated transition region, each include a separate set of pixels that may each be processed separately for suppressing pixel intensities in the estimated background region and improving image quality.

Description

    BACKGROUND OF INVENTION
  • The present disclosure relates generally to a method and product for processing digital images, and particularly to a method and product for processing digital images with detection and suppression of background noise.
  • A digital image is a collection of digital information in the form of pixels that may be processed to provide a visual image representative of an object-of-interest. The digital image may be obtained from digital cameras, digital video, digital scanners, or the like, or may be digitized from an analog film. An exemplary object-of-interest is a biological object, and an exemplary digital scanner is a magnetic resonance imaging (MRI) scanner. While reference may be made herein to MRI as an exemplary scanning method, it will be appreciated that other scanning methods may be employed, such as computed tomography (CT), ultrasound, or X-ray, for example. Digital images of objects that are smaller than the field of view of the scan contain background regions where the signal intensity should be zero. However, variations in the intensities of these background pixels may arise due to noise and other artifacts such as ghosting and gradient warping correction for example. Though noise may be prominent throughout an image, it is most prominent in the background regions where the signal is expected to be zero. The presence of background intensity in an image unnecessarily distracts an observer who may be viewing the final image. Accordingly, there is a need in the art for a digital image processing method and product that improves image quality by detecting and suppressing background noise.
  • SUMMARY OF INVENTION
  • Embodiments of the invention include a method for processing a digital image. A foreground region relating to an imaged object is estimated, a background region relating to other than the imaged object is estimated, and by using the image, the estimated foreground region and the estimated background region, a transition region disposed between the foreground region and the background region is calculated. The estimated foreground region, the estimated background region, and the calculated transition region, each include a separate set of pixels that may each be processed separately for suppressing pixel intensities in the estimated background region and improving image quality.
  • Other embodiments of the invention include a computer program product for processing a digital image, the product including a storage medium, readable by a processing circuit, storing instructions for execution by the processing circuit for performing embodiments of the aforementioned method.
  • BRIEF DESCRIPTION OF DRAWINGS
  • Referring to the exemplary drawings wherein like elements are numbered alike in the accompanying Figures:
  • FIG. 1 depicts an exemplary method and system for processing digital images in accordance with embodiments of the invention;
  • FIG. 2 depicts further detail of the method of FIG. 1;
  • FIG. 3 depicts further detail of the method of FIG. 2;
  • FIG. 4 depicts an alternative method to that depicted in FIG. 3; and
  • FIG. 5 depicts yet further detail of the method of FIG. 1.
  • DETAILED DESCRIPTION
  • Embodiments of the invention provide a method and product for processing digital images, and particularly to a method and product for processing digital images with detection and suppression of background noise. In an exemplary embodiment, a biological object is scanned using an MRI imaging technique, and the resulting image is digitally processed for the detection of background noise and suppression thereof, thereby improving the signal-to-noise-ratio (SNR) of the resulting image. The digital image is separated into foreground, background and transition regions, thereby enabling each region to be analyzed and processed separately. The processing techniques applied involve low-level image processing techniques, such as thresholding and binary image subtraction for example. While embodiments described herein may employ MRI as an exemplary imaging technique, it will be appreciated that the disclosed invention may also employ other imaging techniques, such as CT, Ultrasound, X-ray, and the like, for example.
  • Referring now to FIG. 1, a high-level flowchart of an exemplary method 100 for processing a digital image is depicted. An original image 110 is analyzed and segmented at block 120 into foreground, background and transition regions, with the background region being referred to as a background mask 130, and the foreground and transition regions being incorporated into and referred to as a transition mask 140. A background detection method 200, discussed hereinafter in reference to FIG. 2, segments original image 110 into the three aforementioned regions. While reference is made herein to an “original” image 110, it will be appreciated that this reference is merely a reference to the image entering analysis block 120, and does not infer a “first occurrence” of image 110. For example, it is contemplated that original image 110 may be a digital representation of an analog film resulting from an imaging technique, such as X-ray for example. Here, the analog film would be considered the “first occurrence” image, and the digital representation would be considered the original image 110 entering analysis block 120.
  • Information from original image 110, background mask 130 and transition mask 140, is used by a background noise suppression method 300, discussed hereinafter in reference to FIG. 3, for suppressing the background noise and improving the SNR of the resulting image. The resulting image is then filtered to produce a filtered image 150. While embodiments described herein disclose the suppression of background noise, it will be appreciated that such suppression may generally be described as the suppression of pixel intensities in the background region.
  • Background detection method 200 will now be discussed with reference to FIGS. 2-4. From original image 110, a foreground region relating to the imaged biological object is estimated 202 by applying a pixel intensity threshold t_high 402 to filter out all but those pixels having an intensity equal to or greater than t_high 402, where such filtering may alternatively be viewed as filtering original image 110 according to a first criterion. In an exemplary embodiment, the value of t_high 402 is set to be about 50% of the mean intensity of the non-zero pixels in original image 110. The resulting region is referred to as the estimated foreground region 204, also referred to as the initial foreground region. Also from original image 110, a background region relating to other than the imaged object is estimated as containing all pixels in original image 110 having an intensity of zero. The resulting region is referred to as the estimated background region 206, which is also referred to as the zero mask. Alternatively, the estimated background region 206 may be defined as that region containing those pixels of the image meeting a second criterion, such as an intensity of zero, or an intensity less than a low intensity threshold, for example.
  • From original image 110, a gradient magnitude image 208 is computed that provides a value for the gradient of the intensity of each pixel of original image 110. A gradient-constrained-hysteresis-threshold (GCHT) method 400 is applied to original image 110 using gradient magnitude image 208 to calculate an initial transition region 210. Initial transition region 210 is disposed between estimated foreground and background regions 204, 206, which may be seen by referring to the illustration depicted FIG. 5.
  • In an embodiment, initial transition region 210 is calculated to be that region containing pixels having a morphological connection (that is, a connection made possible via a morphological operation) to a pixel of estimated foreground region 204, having an intensity greater than a low threshold t_low 404, and having a gradient magnitude that is within a gradient tolerance value g_tol 406 of the gradient magnitude of the foreground pixel to which it is connected, which is depicted in algorithm form in FIGS. 3 and 4. Generally, however, initial transition region 210 may be considered to be that region of the image containing those pixels that meet a transition region criterion, such as the set of aforementioned criteria, for example.
  • Referring to FIG. 3, the application of GCHT method 400 with hysteresis thresholding and gradient constraining is depicted in the absence of an iterative approach. Here, the inputs to GCHT method 400 are gradient magnitude image 208 and estimated foreground region 204, where estimated foreground region 204 has been estimated using intensity threshold t_high 402. The output of GCHT method 400 is a gradient constrained hysteresis threshold image 450. For each pixel x in the foreground (estimated foreground region 204 for example) where the pixel intensity Ix is greater than t_high 402, initial transition region 210 is calculated to be that region defined by the set of pixels y where:
      • y is connected to x in an 8-connected sense (that is, in a 3-by-3 matrix of 9 pixels, the center pixel is considered to be connected to each of the 8 pixels on the perimeter);
      • the intensity Iy of y pixel is greater than low threshold t_low 404; and
      • the absolute value of the difference between the gradient magnitude of x pixel and the gradient magnitude of y pixel (|Gx-Gy|) is less than a defined gradient tolerance value g_tol 406.
  • In GCHT 400, the values for low threshold t_low 404 and gradient tolerance value g_tol 406 may be user adjusted, thereby making GCHT method 400 tunable.
  • Estimated foreground region 204, estimated background region 206, and calculated initial transition region 210, each comprise a separate set of pixels that may each be processed separately for suppressing background noise and improving image quality. For example, and as suggested by GCHT method 400 depicted in FIG. 3, the y pixels of initial transition region 210 may be added to the x pixels of estimated foreground region 204, thereby growing the foreground region that is representative of the biological object being imaged for greater object renditioning. Here, and in the absence of an iterative approach, initial transition region 210 becomes transition mask 140.
  • An advantage associated with GCHT method 400 is that in an image having strong edges, the growing region grows along the edge and not across the edge. In this way, moderate intensity background artifacts that are connected to an edge will not get incorporated into the transition region. As a result, the t_low threshold 404 may be set quite low to capture low intensity foreground regions into the transition region without incorporating artifacts with higher intensities that are in the background. Also, this approach will capture the low intensity foreground regions in images that suffer from intensity variation as long as the image gradient in the regions of intensity variation is less than the gradient tolerance value g_tol.
  • Referring to FIG. 4, GCHT method 400 is modified to include an iterative approach, thereby resulting in an iterative GCHT method 500. Here, GCHT method 400 is shown as one operation in an iterative do-loop 505.
  • At the beginning of the iterative process, the gradient tolerance value g_tol 406 is set 510 to a low gradient tolerance value tol_low 408, after which, GCHT method 400 is applied as discussed previously. Next, at 520, it is determined whether the foreground-plus-transition-region (for example, estimated foreground region 204 plus initial transition region 210 for the first pass in the iterative process) has grown by more than a defined number of pixels size_t. In response to the determination at 520 being yes, all pixels that may be added to the foreground since the last labeling are labeled as incremental transition region “i”, and the iteration counter is incremented by one (i=i+1). These actions are depicted at 530. In response to the determination at 520 being no, gradient tolerance value g_tol 406 is incremented by a gradient tolerance value step g_tol_step 410 until a high gradient tolerance value tol_high 412 condition is met or exceeded. These actions are depicted at 540 and 550, respectively, where 550 is labeled “repeat while: g_tol<tol_high”. With each iteration of do-loop 505, incremental transitions regions (i, i+1, i+2, etc.) are calculated, with each incremental transition region having an incrementally larger gradient tolerance value g_tol 406 that has a value between tol_low 408 and tol_high 412. The iterative do-loop 505 is continued until tol_high 412 is met or exceeded.
  • In an embodiment employing the iterative method of FIG. 4, each incremental transition region (i, i+1, i+2, etc.) 530 is calculated to be that region containing pixels connected to a pixel of a previously calculated incremental transition region 530, having an intensity greater than a low threshold t_low 404, and having a gradient magnitude that is within the incrementally larger gradient tolerance value (g_tol+g_tol_step) of the gradient magnitude of the incremental transition region pixel to which it is connected.
  • Upon completion of do-loop 505, that is, where the condition “repeat while:g_tol<tol_high” 550 is no longer satisfied, all of the incremental transition regions (i, i+1, i+2, etc.) 530 are merged into a single transition region 560 using a focus parameter 570. In an embodiment, the focus parameter 570 is the percentage of the total number of transition regions that will be kept as transition mask 140. For example, where there are ten transition regions (i=1 through 10) and the focus parameter 570 is set to 80% (focus=0.8), then the first eight transitions regions (i=1 through 8) will be merged into transition mask 140 and the last two transition regions will be discarded. Accordingly, and in the presence of an iterative approach, the adjustment of single transition region 560 by focus parameter 570 results in transition mask 140. An advantage of using a focus parameter is that the internal parameters of method 100 may be fixed and the entire process may be made controllable or tunable by the single focus parameter 570. The higher the focus parameter 570, the more liberal method 100 is in accepting pixels into transition mask 140.
  • In an embodiment, the following parameters may be tunable:
      • intensity high threshold t_high 402
      • intensity low threshold t_low 404
      • gradient tolerance value g_tol 406
      • low gradient tolerance value tol_low 408
      • gradient tolerance value step g_tol_step 410
      • high gradient tolerance value tol_high 412.
  • However, some of these parameters may set to a defined value. For example, low gradient tolerance value tol_low 408 and high gradient tolerance value tol_high 412 may be set to a specific percentage of the average gradient magnitude of the entire original image 110, and gradient tolerance value step g_tol_step 410 may be set according to a desired number of iterations, such as (tol_high-tol_low)/10 where ten iterations are desired for generating ten incremental transition regions 530, for example.
  • Once transition mask 140 is defined, an object region, also referred to as the object image, may be defined as being the union of the estimated foreground region 204 and the transition mask 140, which may then undergo morphological operations for improving the image quality. Here, the term object region refers to those pixels of the original image that are determined via method 100 to be representative of the biological object under observation. Such morphological operations may include erosion of the image and dilation back to the transition mask to remove small objects and bridges that may connect background noise to the main object, and dilation of the image and erosion back to the original size to fill small cracks in the edges of the image 212. The zero mask may then be combined with the object image, which may then undergo a connected-components morphological operation to fill small holes in the resulting image 214. Upon completion of the hole filling process 214, the zero mask is removed from the image resulting in the final object mask 216. The final object mask 216 is then separated into a final foreground region and a final transition region 218. In accordance with embodiments of the invention disclosed herein, the initial estimated foreground region 204 is labeled the final foreground mask 152, the difference between the object region and the final foreground mask is labeled the final transition mask 154, and the remainder of the image is labeled the final background mask 156, which collectively make up the filtered image 150, best seen by referring back to FIG. 1. By separating the image into different regions, each region may be treated differently to improve image quality. For example, the intensities of the foreground pixels may be increased slightly to improve contrast and increase SNR in the foreground regions, while the intensities of the background pixels are lowered to reduce noise in the background regions. The final object image is the union of the final foreground mask 152 and the final transition mask 154 after filtering.
  • Regarding the suppression of background noise, and referring now to FIG. 5, the background noise may be suppressed gradually by reducing the intensity of the background pixels to zero as a function of their distance from the final object image. In FIG. 5, final foreground mask 152, final transition mask 154, and final background mask 156, are depicted. Shown at increasing distances from final transition mask 154 are two regions 310, 320, which extend into the background region defined by final background mask 156. In accordance with embodiments of the invention, the background noise in regions 310, 320 is suppressed gradually by reducing the intensity of the background pixels in those regions to zero the further the region is from final transition mask 154. Different functions may be employed to gradually reduce the intensity of the background pixels to zero, such as a Gaussian function, a Hanning function, a Hamming function, or any function suitable for reducing a pixel intensity value with respect to distance from a defined origin. Combinations of functions may be employed as desired. Alternatively, all background pixel intensities may be identically set to zero, or the background region may be homogenized by iteratively blurring the background region and adding a small fraction of the original image back into the background region.
  • Embodiments of the invention may be provided in executable instruction form on a storage medium, such as memory 605 or in the form of a hard drive, a floppy disk or a CD-ROM 610 for example, that is readable by a processing circuit, such as processor 615 for example, the processing circuit being in signal communication via application software with a graphical user interface at a computer, such as computer 600 for example, whereby a user may execute the embedded instructions for practicing the disclosed invention. The instructions may be loaded into and/or executed by computer 600, or transmitted over some transmission medium, such as over electrical wiring or cabling, through fiber optics, or via electromagnetic radiation, wherein when the computer program code is loaded into and executed by computer 600, the computer becomes an apparatus for practicing the invention. When implemented on a general-purpose microprocessor, the computer program code segments configure the microprocessor to create specific logic circuits. The technical effect of the executable instructions is to enhance the quality of a digital image, of a scanned biological object for example, by detecting and suppressing background noise in the image.
  • As disclosed, some embodiments of the invention may include some of the following advantages: use of low level image processing tools (no statistical models, no neural network based methods, for example) for improving image quality; use of method parameters that may be tunable or predefined; the ability to capture low intensity foreground regions into the transition mask without incorporating artifacts with higher intensities that are in the background; and, the ability to execute morphological and filtering operations on discrete regions of the image thereby providing greater control over the generation and quality of the final object image.
  • While the invention has been described with reference to exemplary embodiments, it will be understood by those skilled in the art that various changes may be made and equivalents may be substituted for elements thereof without departing from the scope of the invention. In addition, many modifications may be made to adapt a particular situation or material to the teachings of the invention without departing from the essential scope thereof. Therefore, it is intended that the invention not be limited to a particular embodiment disclosed as the best or only mode contemplated for carrying out this invention, but that the invention will include all embodiments falling within the scope of the appended claims. Moreover, the use of the terms first, second, etc. do not denote any order or importance, but rather the terms first, second, etc. are used to distinguish one element from another. Furthermore, the use of the terms a, an, etc. do not denote a limitation of quantity, but rather denote the presence of at least one of the referenced item.

Claims (30)

1. A method for processing a digital image, the method comprising:
estimating a foreground region relating to an imaged object;
estimating a background region relating to other than the imaged object; and
by using the image, the estimated foreground region and the estimated background region, calculating a transition region disposed between the foreground region and the background region;
wherein the estimated foreground region, the estimated background region, and the calculated transition region, each comprise a separate set of pixels that may each be processed separately for suppressing pixel intensities in the estimated background region and improving image quality.
2. The method of claim 1, wherein:
the estimating a foreground region comprises defining an initial foreground region as that region containing those pixels of the image meeting a first criterion; and
the estimating a background region comprises defining the background region as that region containing those pixels of the image meeting a second criterion.
3. The method of claim 2, wherein the first criterion comprises a pixel intensity greater than a first threshold.
4. The method of claim 2, wherein the second criterion comprises a pixel intensity less than a second threshold.
5. The method of claim 2, wherein the calculating a transition region comprises calculating the transition region as that region containing those pixels of the image meeting a third criterion.
6. The method of claim 5, wherein the third criterion comprises:
a pixel having a pixel intensity greater than the second threshold, a morphological connection to a foreground pixel, and a gradient magnitude that is within a gradient tolerance value of the gradient magnitude of the foreground pixel.
7. The method of claim 5, wherein the calculating a transition region further comprises:
iteratively calculating incremental transition regions, each having an incrementally larger gradient tolerance value, until a gradient tolerance value threshold is met or exceeded, wherein:
each incremental transition region is calculated to be that region containing pixels connected to a pixel of a previously calculated incremental transition region, having an intensity greater than the second threshold, and having a gradient magnitude that is within the incrementally larger gradient tolerance value of the gradient magnitude of the incremental transition region pixel to which it is connected.
8. The method of claim 7, wherein the calculating a transition region further comprises:
using a focus parameter to merge a defined number of the initial plus incremental transition regions into a single transition region.
9. The method of claim 8, further comprising:
defining an object region as the union of the initial foreground region and the single transition region, and performing at least one morphological operation on the object region.
10. The method of claim 9, further comprising:
defining a final foreground mask as the initial foreground region;
defining a final transition mask as the difference between the object region and the final foreground region; and
defining a final background mask as the remainder of the image.
11. The method of claim 8, further comprising:
suppressing pixel intensities in the background region by gradually reducing the intensity of background pixels to zero as a function of their distance from the object region.
12. The method of claim 11, wherein the function comprises a linear ramp function, an exponential function, a Gaussian function, a Hanning function, a Hamming function, any function for reducing a value with respect to distance, or any combination of functions comprising at least one of the foregoing functions.
13. The method of claim 5, further comprising:
defining an object region as the union of the initial foreground region and the initial transition region, and performing at least one morphological operation on the object region.
14. The method of claim 13, further comprising:
defining a final foreground mask as the initial foreground region;
defining a final transition mask as the difference between the object region and the final foreground region; and
defining a final background mask as the remainder of the image.
15. The method of claim 14, further comprising:
suppressing pixel intensities in the background region by gradually reducing the intensity of background pixels to zero as a function of their distance from the object region.
16. The method of claim 15, wherein the function comprises a linear ramp function, an exponential function, a Gaussian function, a Hanning function, a Hamming function, any function for reducing a value with respect to distance, or any combination of functions comprising at least one of the foregoing functions.
17. The method of claim 1, wherein the digital image is a digital image of a biological object obtained using MR imaging, CT imaging, Ultrasound imaging, X-ray imaging, or any combination comprising at least one of the foregoing imaging processes.
18. A computer program product for processing a digital image, the product comprising:
a storage medium, readable by a processing circuit, storing instructions for execution by the processing circuit for:
estimating a foreground region relating to an imaged object;
estimating a background region relating to other than the imaged object; and
by using the image, the estimated foreground region and the estimated background region, calculating a transition region disposed between the foreground region and the background region;
wherein the estimated foreground region, the estimated background region, and the calculated transition region, each comprise a separate set of pixels that may each be processed separately for suppressing pixel intensities in the estimated background region and improving image quality.
19. The product of claim 18, wherein:
the estimating a foreground region comprises defining an initial foreground region as that region containing those pixels of the image meeting a first criterion; and
the estimating a background region comprises defining the background region as that region containing those pixels of the image meeting a second criterion.
20. The product of claim 19, wherein the first criterion comprises a pixel intensity greater than a first threshold.
21. The product of claim 19, wherein the second criterion comprises a pixel intensity less than a second threshold.
22. The product of claim 21, wherein the calculating a transition region comprises calculating the transition region as that region containing those pixels of the image meeting a third criterion.
23. The product of claim 22, wherein the third criterion comprises:
a pixel having a pixel intensity greater than the second threshold, a morphological connection to a foreground pixel, and a gradient magnitude that is within a gradient tolerance value of the gradient magnitude of the foreground pixel.
24. The product of claim 23, further comprising instructions for execution by the processing circuit for:
defining an object region as the union of the initial foreground region and the initial transition region, and performing at least one morphological operation on the object region.
25. The product of claim 24, further comprising instructions for execution by the processing circuit for:
defining a final foreground mask as the initial foreground region;
defining a final transition mask as the difference between the object region and the final foreground region; and
defining a final background mask as the remainder of the image.
26. The product of claim 25, further comprising instructions for execution by the processing circuit for:
suppressing pixel intensities in the background region by gradually reducing the intensity of background pixels to zero as a function of their distance from the object region.
27. The product of claim 26, wherein the function comprises a linear ramp function, an exponential function, a Gaussian function, a Hanning function, a Hamming function, any function for reducing a value with respect to distance, or any combination of functions comprising at least one of the foregoing functions.
28. The product of claim 23, wherein the calculating a transition region further comprises:
iteratively calculating incremental transition regions, each having an incrementally larger gradient tolerance value, until a gradient tolerance value threshold is met or exceeded, wherein:
each incremental transition region is calculated to be that region containing pixels connected to a pixel of a previously calculated incremental transition region, having an intensity greater than the second threshold, and having a gradient magnitude that is within the incrementally larger gradient tolerance value of the gradient magnitude of the incremental transition region pixel to which it is connected.
29. The product of claim 28, wherein the calculating a transition region further comprises:
using a focus parameter to merge a defined number of the initial plus incremental transition regions into a single transition region.
30. The product of claim 18, wherein the digital image is a digital image of a biological object obtained using MR imaging, CT imaging, Ultrasound imaging, X-ray imaging, or any combination comprising at least one of the foregoing imaging processes.
US10/708,771 2004-03-24 2004-03-24 Method and product for processing digital images Active 2028-09-24 US7623728B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US10/708,771 US7623728B2 (en) 2004-03-24 2004-03-24 Method and product for processing digital images

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US10/708,771 US7623728B2 (en) 2004-03-24 2004-03-24 Method and product for processing digital images

Publications (2)

Publication Number Publication Date
US20050213845A1 true US20050213845A1 (en) 2005-09-29
US7623728B2 US7623728B2 (en) 2009-11-24

Family

ID=34989873

Family Applications (1)

Application Number Title Priority Date Filing Date
US10/708,771 Active 2028-09-24 US7623728B2 (en) 2004-03-24 2004-03-24 Method and product for processing digital images

Country Status (1)

Country Link
US (1) US7623728B2 (en)

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060210164A1 (en) * 2003-03-04 2006-09-21 Kurokawa Hideyuki Image processing device
US20070053608A1 (en) * 2005-08-23 2007-03-08 Jun Zhang Method for reducing mosquito noise
WO2009048660A3 (en) * 2007-07-16 2009-05-28 Raytheon Co System and method of moving target based calibration of non-uniformity compensation for optical imagers
US20110103692A1 (en) * 2009-10-29 2011-05-05 Raytheon Company Methods and systems for processing data using non-linear slope compensation
WO2011068508A1 (en) * 2009-12-02 2011-06-09 Hewlett-Packard Development Company, Lp System and method of foreground-background segmentation of digitized images
US20110150317A1 (en) * 2009-12-17 2011-06-23 Electronics And Telecommunications Research Institute System and method for automatically measuring antenna characteristics
US20120063656A1 (en) * 2010-09-13 2012-03-15 University Of Southern California Efficient mapping of tissue properties from unregistered data with low signal-to-noise ratio
US8331695B1 (en) * 2009-02-12 2012-12-11 Xilinx, Inc. Integrated circuit having a circuit for and method of updating parameters associated with a background estimation portion of a video frame
WO2013070132A1 (en) * 2011-11-09 2013-05-16 Flir Systems Ab Image processing method for dynamic auto-adjustment of an ir image
US8738678B2 (en) 2009-10-29 2014-05-27 Raytheon Company Methods and systems for determining an enhanced rank order value of a data set
US8786873B2 (en) 2009-07-20 2014-07-22 General Electric Company Application server for use with a modular imaging system
US20140241589A1 (en) * 2011-06-17 2014-08-28 Daniel Weber Method and apparatus for the detection of visibility impairment of a pane
WO2016172527A3 (en) * 2015-04-23 2017-01-19 Bd Kiestra B.V. Colony contrast gathering
US20190206117A1 (en) * 2017-12-29 2019-07-04 UBTECH Robotics Corp. Image processing method, intelligent terminal, and storage device
US20210142455A1 (en) * 2018-08-01 2021-05-13 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Method and apparatus for image processing
WO2024069547A1 (en) * 2022-09-29 2024-04-04 Novocure Gmbh Removal of background noise from image

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7974485B1 (en) * 2005-10-27 2011-07-05 Nvidia Corporation Split-frame post-processing in a programmable video pipeline
JP4245045B2 (en) * 2006-12-26 2009-03-25 ソニー株式会社 Imaging apparatus, imaging signal processing method, and program
US8116585B2 (en) * 2007-08-09 2012-02-14 Xerox Corporation Background noise detection on rendered documents
KR101348596B1 (en) * 2008-01-22 2014-01-08 삼성전자주식회사 Apparatus and method for immersive generation
CN101587586B (en) * 2008-05-20 2013-07-24 株式会社理光 Device and method for processing images
US8457440B1 (en) * 2009-01-27 2013-06-04 Axsun Technologies, Inc. Method and system for background subtraction in medical optical coherence tomography system
US8243882B2 (en) 2010-05-07 2012-08-14 General Electric Company System and method for indicating association between autonomous detector and imaging subsystem
CN103959335B (en) * 2011-11-28 2018-03-09 皇家飞利浦有限公司 Image processing apparatus
JP6687524B2 (en) 2014-01-30 2020-04-22 ビーディー キエストラ ベスローテン フェンノートシャップ System and method for image acquisition using supervised high quality imaging
US11800056B2 (en) 2021-02-11 2023-10-24 Logitech Europe S.A. Smart webcam system
US11800048B2 (en) 2021-02-24 2023-10-24 Logitech Europe S.A. Image generating system with background replacement or modification capabilities

Citations (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4951140A (en) * 1988-02-22 1990-08-21 Kabushiki Kaisha Toshiba Image encoding apparatus
US5157740A (en) * 1991-02-07 1992-10-20 Unisys Corporation Method for background suppression in an image data processing system
US5268967A (en) * 1992-06-29 1993-12-07 Eastman Kodak Company Method for automatic foreground and background detection in digital radiographic images
US5368033A (en) * 1993-04-20 1994-11-29 North American Philips Corporation Magnetic resonance angiography method and apparatus employing an integration projection
US5631975A (en) * 1992-04-14 1997-05-20 Koninkl Philips Electronics Nv Image segmentation device
US5694478A (en) * 1994-12-15 1997-12-02 Minnesota Mining And Manufacturing Company Method and apparatus for detecting and identifying microbial colonies
US5825910A (en) * 1993-12-30 1998-10-20 Philips Electronics North America Corp. Automatic segmentation and skinline detection in digital mammograms
US5915044A (en) * 1995-09-29 1999-06-22 Intel Corporation Encoding video images using foreground/background segmentation
US6061476A (en) * 1997-11-24 2000-05-09 Cognex Corporation Method and apparatus using image subtraction and dynamic thresholding
US6081626A (en) * 1995-05-09 2000-06-27 International Business Machines Corporation Method and system for background removal in electronically scanned images
US6088392A (en) * 1997-05-30 2000-07-11 Lucent Technologies Inc. Bit rate coder for differential quantization
US6173083B1 (en) * 1998-04-14 2001-01-09 General Electric Company Method and apparatus for analyzing image structures
US6240215B1 (en) * 1998-09-23 2001-05-29 Xerox Corporation Method and apparatus for digital image processing with selectable background suppression data acquisition modes
US6243070B1 (en) * 1998-10-07 2001-06-05 Microsoft Corporation Method and apparatus for detecting and reducing color artifacts in images
US6275304B1 (en) * 1998-12-22 2001-08-14 Xerox Corporation Automated enhancement of print quality based on feature size, shape, orientation, and color
US20010055421A1 (en) * 1997-02-10 2001-12-27 Martin Baatz Method of iterative segmentation of a digital picture
US6337925B1 (en) * 2000-05-08 2002-01-08 Adobe Systems Incorporated Method for determining a border in a complex scene with applications to image masking
US20020037103A1 (en) * 2000-01-06 2002-03-28 Hong Qi He Method of and apparatus for segmenting a pixellated image
US6453069B1 (en) * 1996-11-20 2002-09-17 Canon Kabushiki Kaisha Method of extracting image from input image using reference image
US6507618B1 (en) * 2000-04-25 2003-01-14 Hewlett-Packard Company Compressed video signal including independently coded regions
US20030044045A1 (en) * 2001-06-04 2003-03-06 University Of Washington Video object tracking by estimating and subtracting background
US6580812B1 (en) * 1998-12-21 2003-06-17 Xerox Corporation Methods and systems for automatically adding motion lines representing motion to a still image
US20030152285A1 (en) * 2002-02-03 2003-08-14 Ingo Feldmann Method of real-time recognition and compensation of deviations in the illumination in digital color images
US6661918B1 (en) * 1998-12-04 2003-12-09 Interval Research Corporation Background estimation and segmentation based on range and color
US20050055658A1 (en) * 2003-09-09 2005-03-10 International Business Machines Corporation Method for adaptive segment refinement in optical proximity correction
US7391895B2 (en) * 2003-07-24 2008-06-24 Carestream Health, Inc. Method of segmenting a radiographic image into diagnostically relevant and diagnostically irrelevant regions

Patent Citations (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4951140A (en) * 1988-02-22 1990-08-21 Kabushiki Kaisha Toshiba Image encoding apparatus
US5157740A (en) * 1991-02-07 1992-10-20 Unisys Corporation Method for background suppression in an image data processing system
US5631975A (en) * 1992-04-14 1997-05-20 Koninkl Philips Electronics Nv Image segmentation device
US5268967A (en) * 1992-06-29 1993-12-07 Eastman Kodak Company Method for automatic foreground and background detection in digital radiographic images
US5368033A (en) * 1993-04-20 1994-11-29 North American Philips Corporation Magnetic resonance angiography method and apparatus employing an integration projection
US5825910A (en) * 1993-12-30 1998-10-20 Philips Electronics North America Corp. Automatic segmentation and skinline detection in digital mammograms
US5694478A (en) * 1994-12-15 1997-12-02 Minnesota Mining And Manufacturing Company Method and apparatus for detecting and identifying microbial colonies
US6081626A (en) * 1995-05-09 2000-06-27 International Business Machines Corporation Method and system for background removal in electronically scanned images
US5915044A (en) * 1995-09-29 1999-06-22 Intel Corporation Encoding video images using foreground/background segmentation
US6453069B1 (en) * 1996-11-20 2002-09-17 Canon Kabushiki Kaisha Method of extracting image from input image using reference image
US20010055421A1 (en) * 1997-02-10 2001-12-27 Martin Baatz Method of iterative segmentation of a digital picture
US6088392A (en) * 1997-05-30 2000-07-11 Lucent Technologies Inc. Bit rate coder for differential quantization
US6061476A (en) * 1997-11-24 2000-05-09 Cognex Corporation Method and apparatus using image subtraction and dynamic thresholding
US6173083B1 (en) * 1998-04-14 2001-01-09 General Electric Company Method and apparatus for analyzing image structures
US6240215B1 (en) * 1998-09-23 2001-05-29 Xerox Corporation Method and apparatus for digital image processing with selectable background suppression data acquisition modes
US6243070B1 (en) * 1998-10-07 2001-06-05 Microsoft Corporation Method and apparatus for detecting and reducing color artifacts in images
US6661918B1 (en) * 1998-12-04 2003-12-09 Interval Research Corporation Background estimation and segmentation based on range and color
US6580812B1 (en) * 1998-12-21 2003-06-17 Xerox Corporation Methods and systems for automatically adding motion lines representing motion to a still image
US6275304B1 (en) * 1998-12-22 2001-08-14 Xerox Corporation Automated enhancement of print quality based on feature size, shape, orientation, and color
US20020037103A1 (en) * 2000-01-06 2002-03-28 Hong Qi He Method of and apparatus for segmenting a pixellated image
US6507618B1 (en) * 2000-04-25 2003-01-14 Hewlett-Packard Company Compressed video signal including independently coded regions
US6337925B1 (en) * 2000-05-08 2002-01-08 Adobe Systems Incorporated Method for determining a border in a complex scene with applications to image masking
US20030044045A1 (en) * 2001-06-04 2003-03-06 University Of Washington Video object tracking by estimating and subtracting background
US20030152285A1 (en) * 2002-02-03 2003-08-14 Ingo Feldmann Method of real-time recognition and compensation of deviations in the illumination in digital color images
US7391895B2 (en) * 2003-07-24 2008-06-24 Carestream Health, Inc. Method of segmenting a radiographic image into diagnostically relevant and diagnostically irrelevant regions
US20050055658A1 (en) * 2003-09-09 2005-03-10 International Business Machines Corporation Method for adaptive segment refinement in optical proximity correction

Cited By (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060210164A1 (en) * 2003-03-04 2006-09-21 Kurokawa Hideyuki Image processing device
US7899271B1 (en) 2004-09-15 2011-03-01 Raytheon Company System and method of moving target based calibration of non-uniformity compensation for optical imagers
US20070053608A1 (en) * 2005-08-23 2007-03-08 Jun Zhang Method for reducing mosquito noise
US7734089B2 (en) * 2005-08-23 2010-06-08 Trident Microsystems (Far East) Ltd. Method for reducing mosquito noise
WO2009048660A3 (en) * 2007-07-16 2009-05-28 Raytheon Co System and method of moving target based calibration of non-uniformity compensation for optical imagers
US8331695B1 (en) * 2009-02-12 2012-12-11 Xilinx, Inc. Integrated circuit having a circuit for and method of updating parameters associated with a background estimation portion of a video frame
US8786873B2 (en) 2009-07-20 2014-07-22 General Electric Company Application server for use with a modular imaging system
US8416986B2 (en) * 2009-10-29 2013-04-09 Raytheon Company Methods and systems for processing data using non-linear slope compensation
US20110103692A1 (en) * 2009-10-29 2011-05-05 Raytheon Company Methods and systems for processing data using non-linear slope compensation
US8738678B2 (en) 2009-10-29 2014-05-27 Raytheon Company Methods and systems for determining an enhanced rank order value of a data set
WO2011068508A1 (en) * 2009-12-02 2011-06-09 Hewlett-Packard Development Company, Lp System and method of foreground-background segmentation of digitized images
US8792711B2 (en) 2009-12-02 2014-07-29 Hewlett-Packard Development Company, L.P. System and method of foreground-background segmentation of digitized images
US20110150317A1 (en) * 2009-12-17 2011-06-23 Electronics And Telecommunications Research Institute System and method for automatically measuring antenna characteristics
US20120063656A1 (en) * 2010-09-13 2012-03-15 University Of Southern California Efficient mapping of tissue properties from unregistered data with low signal-to-noise ratio
US20140241589A1 (en) * 2011-06-17 2014-08-28 Daniel Weber Method and apparatus for the detection of visibility impairment of a pane
WO2013070132A1 (en) * 2011-11-09 2013-05-16 Flir Systems Ab Image processing method for dynamic auto-adjustment of an ir image
US11341648B2 (en) 2015-04-23 2022-05-24 Timothy M. Wiles Colony contrast gathering
WO2016172527A3 (en) * 2015-04-23 2017-01-19 Bd Kiestra B.V. Colony contrast gathering
US10521910B2 (en) 2015-04-23 2019-12-31 Bd Kiestra B.V. Colony contrast gathering
US10692216B2 (en) 2015-04-23 2020-06-23 Bd Kiestra B.V. Colony contrast gathering
US11669971B2 (en) 2015-04-23 2023-06-06 Bd Kiestra B.V. Colony contrast gathering
US20190206117A1 (en) * 2017-12-29 2019-07-04 UBTECH Robotics Corp. Image processing method, intelligent terminal, and storage device
US20210142455A1 (en) * 2018-08-01 2021-05-13 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Method and apparatus for image processing
US11783462B2 (en) * 2018-08-01 2023-10-10 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Method and apparatus for image processing
WO2024069547A1 (en) * 2022-09-29 2024-04-04 Novocure Gmbh Removal of background noise from image

Also Published As

Publication number Publication date
US7623728B2 (en) 2009-11-24

Similar Documents

Publication Publication Date Title
US7623728B2 (en) Method and product for processing digital images
US7835562B2 (en) Methods and apparatus for noise reduction filtering of images
EP2380132B1 (en) Denoising medical images
Ju et al. Gamma-correction-based visibility restoration for single hazy images
JP5645399B2 (en) X-ray image processing apparatus, X-ray image processing method, and computer program
Çalişkan et al. An efficient noisy pixels detection model for CT images using extreme learning machines
JP2005228342A (en) Method and system for segmenting scanned document
EP2869261A1 (en) Method for processing image data representing a three-dimensional volume
Yang et al. Antialiasing recovery
WO2009030596A1 (en) Method for non-photorealistic rendering
HosseinKhani et al. Adaptive real-time removal of impulse noise in medical images
Shedbalkar et al. A comparative analysis of filters for noise reduction and smoothening of brain MRI images
KR20140109801A (en) Method and apparatus for enhancing quality of 3D image
Tao et al. Adaptive image enhancement method using contrast limitation based on multiple layers BOHE
US9153016B2 (en) Method and system for enhancing contrast of spatially-localized phenomena in mammography image
Zhan et al. Nonlocal means image denoising with minimum MSE-based decay parameter adaptation
US8538163B2 (en) Method and system for detecting edges within an image
US20060008174A1 (en) Count adaptive noise reduction method of x-ray images
CN114299081B (en) Maxillary sinus CBCT image segmentation method, maxillary sinus CBCT image segmentation device, maxillary sinus CBCT storage medium and electronic equipment
US20100061656A1 (en) Noise reduction of an image signal
CN116228553A (en) Image enhancement method capable of simultaneously enhancing definition of high-illumination and low-illumination areas
KR20190067551A (en) Method for contrast enhancement of medical diagnostic image using statistical estimation and mapping function modeling technique
Chandran et al. Segmentation of dental radiograph images
Dasgupta Comparative analysis of non-blind deblurring methods for noisy blurred images
JP7159341B2 (en) Learning-Based Voxel Evolution for Regularized Reconstruction

Legal Events

Date Code Title Description
AS Assignment

Owner name: GENERAL ELECTRIC COMPANY, NEW YORK

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:AVINASH, GOPAL B.;LAL, RAKESH MOHAN;REEL/FRAME:014444/0451

Effective date: 20040317

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCF Information on status: patent grant

Free format text: PATENTED CASE

CC Certificate of correction
FPAY Fee payment

Year of fee payment: 4

FPAY Fee payment

Year of fee payment: 8

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 12