US20050093890A1 - System and method for displaying images utilizing multi-blending - Google Patents

System and method for displaying images utilizing multi-blending Download PDF

Info

Publication number
US20050093890A1
US20050093890A1 US10/859,747 US85974704A US2005093890A1 US 20050093890 A1 US20050093890 A1 US 20050093890A1 US 85974704 A US85974704 A US 85974704A US 2005093890 A1 US2005093890 A1 US 2005093890A1
Authority
US
United States
Prior art keywords
recited
image
images
feature class
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US10/859,747
Other versions
US7667717B2 (en
Inventor
Patrick Baudisch
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Microsoft Technology Licensing LLC
Original Assignee
Microsoft Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Microsoft Corp filed Critical Microsoft Corp
Priority to KR1020040040296A priority Critical patent/KR101076900B1/en
Priority to US10/859,747 priority patent/US7667717B2/en
Assigned to MICROSOFT CORPORATION reassignment MICROSOFT CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BAUDISCH, PATRICK M.
Publication of US20050093890A1 publication Critical patent/US20050093890A1/en
Application granted granted Critical
Publication of US7667717B2 publication Critical patent/US7667717B2/en
Assigned to MICROSOFT TECHNOLOGY LICENSING, LLC reassignment MICROSOFT TECHNOLOGY LICENSING, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MICROSOFT CORPORATION
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/50Lighting effects
    • G06T15/503Blending, e.g. for anti-aliasing
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G5/00Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
    • G09G5/14Display of multiple viewports
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2340/00Aspects of display data processing
    • G09G2340/10Mixing of images, i.e. displayed pixel being the result of an operation, e.g. adding, on the corresponding input pixels
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2340/00Aspects of display data processing
    • G09G2340/14Solving problems related to the presentation of information to be displayed
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G5/00Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
    • G09G5/02Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the way in which colour is displayed
    • G09G5/026Control of mixing and/or overlay of colours in general

Definitions

  • the present application relates to computer software, and in particular, to a system and method for displaying multiple images on a graphical user interface utilizing multi-blending.
  • ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇
  • each instantiated software application can be displayed to the user in the graphical display as a graphical window.
  • each graphical window can be organized in manner that allows multiple graphical windows to be visible to the user at the same time. Because multiple graphical windows may be displayed concurrently to the user, there are situations in which at least a portion of the graphical windows overlaps. Because two or more graphical windows may be overlapping, a typical operating environment resolves an overlap in a variety of manners.
  • each graphical window is associated with an order display priority.
  • the window with the greatest display priority will be displayed in its entirety.
  • Any underlying graphical windows with a lower priority are displayed with the overlapping portion of the graphical window removed.
  • an illustrative screen display 100 includes two graphical windows 102 , 104 corresponding to instantiated software applications. As illustrated in FIG. 1 , at least a portion of the graphical windows 102 , 104 overlaps, as illustrated at 106 . Assuming that the graphical window 104 has a higher display priority, graphical window 104 is display in its entirety, while graphical window 102 is displayed with the overlapping portion 106 omitted. Thus, graphical window 104 appears to be the foreground image, while graphical window 102 appears to become the background image.
  • a software application may be associated with one or more tools that are displayed to the user as part of the display of the software application.
  • the user tools are often referred to as “palettes.” Palettes can be displayed to the user as separate graphical windows that have a higher display priority than the underlying software application graphical window.
  • FIG. 2 an illustrative screen display 200 includes a single graphical display corresponding to an instantiated software application.
  • the screen display 200 includes some underlying content 202 being displayed to the user.
  • two palettes 206 , 208 are displayed with a higher display order than the underlying screen display 200 .
  • At least a portion of the underlying content from a graphical window having a lower display priority is occluded from the view of the user.
  • users can adjust location of each graphical window and/or utilize larger screen displays or multiple screen displays, there are many embodiments in which graphical windows may overlap and in which the user does not wish to have the underlying content complete occluded from view.
  • alpha blending relates to the association of a single weighted value to the color value data for each pixel of a foreground and background image. The degree of transparency corresponds to the weight placed on the foreground image.
  • alpha-blending techniques can become deficient in a variety of manners.
  • the blending of foreground and background images can affect the readability of the either the foreground and background content.
  • conventional alpha blending techniques may make it difficult for a user to determine whether the “blended” content belongs to the foreground or background image.
  • a system and method for processing images utilizing varied feature class weights is provided.
  • a computer system associates two or more images with a set of feature class data, such as color and texture data.
  • the computer system assigns a set of processing weights for each of the feature classes.
  • the two or more images are blended according to the feature class weights.
  • the blended image can be further adjusted according to the content of the images.
  • a method for processing two or more images represented by a set of feature classes is provided.
  • a computer system associates each of the two or more images with a set of feature class data.
  • the computer system assigns a processing weight for each feature class.
  • the computer system processes the set of feature class data according to the assigned processing weight to generate a blended image.
  • a method for blending a foreground image and a background image is provided.
  • the foreground and background images are represented by a set of feature classes.
  • a computer system assigns a processing weight for each feature class.
  • the computer system then processes the set of feature class data for the foreground and background image according to the assigned processing weight to generate a blended image. Additionally, the computer system adjusts the blended image according to the content of the foreground and background image.
  • a method for processing two or more images represented by a set of feature classes is provided.
  • a computer system associates each of the two or more images with a set of feature class data.
  • the computer system assigns a processing weight for each feature class and processes the set of feature class data according to the assigned processing weight to generate a blended image.
  • the computer system then adjusts the blended image according to the content of the two or more images.
  • FIG. 1 is a block diagram of a screen display illustrating overlapping graphical windows in accordance with a conventional display embodiment
  • FIG. 2 is a block diagram of a screen display illustrating overlapping palette graphical windows in accordance with a conventional display embodiment
  • FIG. 3 is a flow diagram illustrative of a multi-blending display routine implemented by a computer system in accordance with an embodiment of the present invention
  • FIG. 4 is a flow diagram illustrative of a palette generation routine utilizing multi-blending and implemented by a computer in accordance with an embodiment of the present invention.
  • FIG. 5 is a block diagram illustrative of a screen display illustrating the processing of images utilizing multi-blending.
  • the present invention relates to a system and method for processing images. More particularly, the present invention relates to a system and method for applying individual weights to a set of features representative of two or more blended images.
  • Features generally correspond to one or more pieces of data describing an attribute of an image and utilized by a computer system to render the image on a display.
  • feature data can include color class data, such as RGB and CIE Lab color model.
  • RGB color class data
  • CIE Lab color model corresponds to a perception-oriented color model that represents pixel color values in three channels, namely, luminance “L”, red-green color different “A”, and blue-yellow color difference “B”.
  • the CIE Lab color model can be further described in K.
  • feature data can also include texture class data that allows users to perceive contrast between images, such as edges or other areas of high contrast.
  • pixel display attributes may be expressed in a CIE Lab color model according to “L”, “a”, and “b” values.
  • the overlapping content can be considered a weighted combination of the individual “L”, “a”, “b” color values for the two or more overlapping images.
  • the overlapping content can also be represented according to a weighted combination of the texture values of both images.
  • the weights applied to each feature may be different.
  • ⁇ 1, ⁇ 2, and ⁇ 3, are selected in a manner that facilitates user recognition of the perception of each image.
  • the individual weight values may correspond to a range of numbers from 0 to 1.
  • each weight value corresponds to a Boolean value of either “1” or “0”.
  • each feature class will correspond only to the channel values from one of the overlapping images.
  • Equation 2 may be modified to include texture class data by the inclusion of a weight value or each texture class channel.
  • Equation 3 corresponds to the modification of Equation 2 to include one or more channels of texture class data.
  • LxAxBx+ ⁇ 1 ⁇ tX— ( ⁇ 1 L 1+(1 ⁇ 1) L 2)+( ⁇ 2 a 1+(1 ⁇ 2) a 2)+( ⁇ 3 b 1+(1 ⁇ 3) b 2)+( ⁇ 4 ⁇ 1 1+(1 ⁇ 4) ⁇ 2 1) . . . +( ⁇ X ⁇ 1 X+ (1 ⁇ X ) ⁇ 2 X ) (3) where:
  • the individual channel weights may be pre-assigned by a user, by the software application, or according to the content being rendered.
  • the individual channel weights may relate to the type of overlapping content (e.g., overlapping palette graphical windows) to achieve a specific effect.
  • An example of the blending of overlapping palette graphical windows will be described in greater detail below.
  • the channel weight values may be expressed as a series of weight values that may be manually or dynamically modified.
  • a computer may utilize a series of channel weights to achieve a transition between blended images.
  • a routine 300 for blending multiple images will be described.
  • a computer system associates two or more images to be blended with a set of feature classes.
  • each pixel in an image may be represented according to its L, a and b CIE Lab color model value. Additionally, each pixel in an image may also be represented by one or more texture values as well.
  • the routine 300 can include the translation of pixel color model data from another other color model to the CIE Lab color model.
  • the computer system associates channel weights for each of the feature classes.
  • the channel weights may be pre-assigned based on user input, computer settings and/or the type of windows being blended. For example, a graphical window corresponding to a palette may be blended with different channel weights than two overlapping graphical windows corresponding to instantiated software applications.
  • the individual channel weights can be expressed as Boolean values.
  • the two images are processed according to the particular channel weights to generate a combined set of feature values for each pixel.
  • the image pixel data may be processed in a manner corresponding to Equation 2. For example, if the channel weights ⁇ 1, ⁇ 2, ⁇ 3, are set to “1”, “0” and “1” values, the resulting pixel image would utilize the L and b values from the first image and the “a” color value from the second image. Additionally, if the texture channel weight is set to “0”, the texture values, e.g., the edge values from the second image will be used.
  • the net set of feature values may be adjusted to improve image displays.
  • image processing filters such as contrast filters, color filters, blur filters and the like may be utilized to adjust the foreground or background images.
  • channel data from one of the images may be ported to another channel to distinguish the foreground and background images.
  • usage data may be utilized to diminish portions of a foreground image in favor of underlying content associated with a background image.
  • other traditional image processing techniques such as alpha blending may also be utilized to improve the image displays.
  • FIG. 4 is a flow diagram illustrative of a palette multi-blending routine 400 implemented by a computer system in accordance with the present invention.
  • Routine 400 corresponds to block 308 in which color channel weight values have been set for two images in which the background image corresponds to a photograph or other content and the foreground image corresponds to a palette. Based on the characteristics of the content, channel weights ⁇ 1, ⁇ 2, ⁇ 3, are set to “1”, “0” and “0”. Additionally, the texture values are set to “1.”
  • the images are desaturated to apply the color channel weights.
  • the blending image includes the red/green and blue/yellow values from the background image.
  • the luminance values correspond solely to the foreground image.
  • the background image colors are preserved and not diluted.
  • the palette surfaces are made transparent to apply the texture channel settings.
  • palettes typically consist of icons corresponding to controls, whose contours can be recognized by users.
  • the image is filtered to with a high-pass filter, such as an emboss filter, to bring out the edges of the controls. In essence, the image is filtered to call out its texture channel values.
  • additional blending techniques such as the linear light blending technique, may be applied to preserve light/dark contrast.
  • blocks 402 and 404 correspond to the application of Boolean weight channels.
  • the routine 400 can include additional processes that improve the overall improvement of the blended image (block 308 , FIG. 3 ).
  • the image improvement processes may be optional and/or dependent on the type of image content being blended. For example, the blending of two graphic images may require a different set of improvement processes than the blending of two textual messages.
  • a portion of the background image is blurred to eliminate high frequency interference associated with noisy backgrounds. More specifically, the overlapping portion of the background image may be blurred with a blur filter.
  • the blur filter may be configured to dynamically blur any portion of the background image overlapping with the foreground image as the foreground image moves.
  • the resulting image provides high contrast for the foreground image and low contrast for the overlapping portion of the background image.
  • usage data may be utilized to vary opacity of portions of the foreground image.
  • usage data corresponding to selection data or other explicit or implicit user feedback systems is used to vary portions of the display. For example, borders, unused icons, or persistent icons that a user is familiar with from a foreground image, such as a palette, may be diminished in favor of the background image.
  • one or more feature channels may be remapped in the event of conflicting images.
  • the foreground and background image may correspond to two images utilizing similar features to represent the content.
  • two images corresponding primarily to text may rely heavily on the luminance channel to represent.
  • only one of the images will be able to use its luminance values because they otherwise interfere.
  • one of the image's luminance values may be mapped to another channel, such as the red/green and blue/yellow difference values to mitigate interference.
  • additional adjustment processes may also be utilized.
  • the routine 400 terminates.
  • FIG. 5 is a block diagram illustrative of a portion of a screen display 500 illustrating the results of processing a foreground 502 and background image 504 in accordance with the present invention.
  • user recognition of the background image 504 is enhanced by preserving color channel values.
  • foreground image 502 recognition is achieved by preserving luminance and texture channel values.
  • the resulting screen display 500 is illustrative in nature and that alternative results may also be achieved within the scope of the present invention.

Abstract

A system and method for processing images utilizing varied feature class weights is provided. A computer system associates two or more images with a set of feature class data, such as color and texture data. The computer system assigns a set of processing weights for each of the feature classes. The two or more images are blended according to the feature class weights. The blended image can be further adjusted according to the content of the images.

Description

    CROSS-REFERENCES TO RELATED APPLICATIONS
  • This application claims the benefit of U.S. Provisional Application No. 60/475,192, entitled “MERGING WINDOWS: TECHNIQUES FOR SIMULTANEOUSLY DISPLAYING TWO OR MORE WINDOWS ON THE SAME SCREEN SPACE” and filed on Jun. 3, 2003. This application also claims the benefit of U.S. Provisional Application No. 60/478,294, entitled “MERGING WINDOWS: TECHNIQUES FOR SIMULTANEOUSLY DISPLAYING TWO OR MORE WINDOWS ON THE SAME SCREEN SPACE,” and filed on Jun. 12, 2003. U.S. Provisional Application Nos. 60/475,192 and 60/478,294 are incorporated by reference herein.
  • FIELD OF THE INVENTION
  • In general, the present application relates to computer software, and in particular, to a system and method for displaying multiple images on a graphical user interface utilizing multi-blending.
  • BACKGROUND OF THE INVENTION
  • Generally described, computer systems provide users with an opportunity to use a number of software applications. In many operating environments, users can instantiate two or more software applications at the same time, often referred to as multi-tasking. For example, a user can instantiate a word processing program, electronic mail program and Internet browser software application at the same. To access each instantiated software application, the user can manipulate various controls, such as a mouse or keyboard, to select a particular instantiated software application.
  • In a typical embodiment, at least of portion of each instantiated software application can be displayed to the user in the graphical display as a graphical window. Additionally, each graphical window can be organized in manner that allows multiple graphical windows to be visible to the user at the same time. Because multiple graphical windows may be displayed concurrently to the user, there are situations in which at least a portion of the graphical windows overlaps. Because two or more graphical windows may be overlapping, a typical operating environment resolves an overlap in a variety of manners.
  • In the simplest solution, each graphical window is associated with an order display priority. In the event that graphical windows overlap, the window with the greatest display priority will be displayed in its entirety. Any underlying graphical windows with a lower priority are displayed with the overlapping portion of the graphical window removed. With reference to FIG. 1, an illustrative screen display 100 includes two graphical windows 102, 104 corresponding to instantiated software applications. As illustrated in FIG. 1, at least a portion of the graphical windows 102, 104 overlaps, as illustrated at 106. Assuming that the graphical window 104 has a higher display priority, graphical window 104 is display in its entirety, while graphical window 102 is displayed with the overlapping portion 106 omitted. Thus, graphical window 104 appears to be the foreground image, while graphical window 102 appears to become the background image.
  • In addition to resolving display priorities between graphical windows corresponding to instantiated software applications, in another typical embodiment, a software application may be associated with one or more tools that are displayed to the user as part of the display of the software application. The user tools are often referred to as “palettes.” Palettes can be displayed to the user as separate graphical windows that have a higher display priority than the underlying software application graphical window. With reference now to FIG. 2, an illustrative screen display 200 includes a single graphical display corresponding to an instantiated software application. The screen display 200 includes some underlying content 202 being displayed to the user. In addition to the screen display 200, two palettes 206, 208 are displayed with a higher display order than the underlying screen display 200.
  • In both of the above examples, at least a portion of the underlying content from a graphical window having a lower display priority is occluded from the view of the user. Although users can adjust location of each graphical window and/or utilize larger screen displays or multiple screen displays, there are many embodiments in which graphical windows may overlap and in which the user does not wish to have the underlying content complete occluded from view.
  • One approach for allowing at least a portion of underlying content to be visible to a user corresponds to the association of a transparency property to the higher display priority graphical window. A typical conventional approach to associating transparency properties to images is referred to as alpha blending. One skilled in the relevant art will appreciate that alpha blending relates to the association of a single weighted value to the color value data for each pixel of a foreground and background image. The degree of transparency corresponds to the weight placed on the foreground image. Equation 1 defines alpha blending for a color schema defining images according to its red, green and blue color values (“RGB”) as:
    RxGxBx=R1α+R2(1−α)+G1α+G2(1−α)+B1α+B2(1−α)  (1)
    where:
      • R1G1B1=color pixel values for a first image;
      • R2G2B2=color pixel values for a second image; and
      • α=alpha value
        The utilization of RGB values to represent images and the utilization of alpha blending to provide transparency properties to images are well known in the art and will not be described in greater detail.
  • Although the utilization of alpha blending facilitates the display of underlying content, alpha-blending techniques can become deficient in a variety of manners. In one aspect, the blending of foreground and background images can affect the readability of the either the foreground and background content. In another aspect, in the event that content from both the foreground and background is similar in some way (such as color, shape, size, etc.) conventional alpha blending techniques may make it difficult for a user to determine whether the “blended” content belongs to the foreground or background image.
  • Thus, there is a need for a system and method for displaying images that facilitates content readability and/or content identification.
  • SUMMARY OF THE INVENTION
  • A system and method for processing images utilizing varied feature class weights is provided. A computer system associates two or more images with a set of feature class data, such as color and texture data. The computer system assigns a set of processing weights for each of the feature classes. The two or more images are blended according to the feature class weights. The blended image can be further adjusted according to the content of the images.
  • In accordance with an aspect of the present invention, a method for processing two or more images represented by a set of feature classes is provided. In accordance with the method, a computer system associates each of the two or more images with a set of feature class data. The computer system assigns a processing weight for each feature class. Additionally, the computer system processes the set of feature class data according to the assigned processing weight to generate a blended image.
  • In accordance with another aspect of the present invention, a method for blending a foreground image and a background image is provided. The foreground and background images are represented by a set of feature classes. In accordance with the method, a computer system assigns a processing weight for each feature class. The computer system then processes the set of feature class data for the foreground and background image according to the assigned processing weight to generate a blended image. Additionally, the computer system adjusts the blended image according to the content of the foreground and background image.
  • In accordance with still a further aspect of the present invention a method for processing two or more images represented by a set of feature classes is provided. A computer system associates each of the two or more images with a set of feature class data. The computer system assigns a processing weight for each feature class and processes the set of feature class data according to the assigned processing weight to generate a blended image. The computer system then adjusts the blended image according to the content of the two or more images.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The foregoing aspects and many of the attendant advantages of this invention will become more readily appreciated as the same become better understood by reference to the following detailed description, when taken in conjunction with the accompanying drawings, wherein:
  • FIG. 1 is a block diagram of a screen display illustrating overlapping graphical windows in accordance with a conventional display embodiment;
  • FIG. 2 is a block diagram of a screen display illustrating overlapping palette graphical windows in accordance with a conventional display embodiment;
  • FIG. 3 is a flow diagram illustrative of a multi-blending display routine implemented by a computer system in accordance with an embodiment of the present invention;
  • FIG. 4 is a flow diagram illustrative of a palette generation routine utilizing multi-blending and implemented by a computer in accordance with an embodiment of the present invention; and
  • FIG. 5 is a block diagram illustrative of a screen display illustrating the processing of images utilizing multi-blending.
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT
  • Generally described, the present invention relates to a system and method for processing images. More particularly, the present invention relates to a system and method for applying individual weights to a set of features representative of two or more blended images. Features generally correspond to one or more pieces of data describing an attribute of an image and utilized by a computer system to render the image on a display. In an illustrative embodiment of the present invention, feature data can include color class data, such as RGB and CIE Lab color model. One skilled in the relevant art will appreciate that the CIE Lab color model corresponds to a perception-oriented color model that represents pixel color values in three channels, namely, luminance “L”, red-green color different “A”, and blue-yellow color difference “B”. The CIE Lab color model can be further described in K. McLaren, “The development of the CIE 1976 (L*a*b*) uniform colour-space and colour-difference formula,” Journal of the Society of Dyers and Colourists, 92, 1976, 338-341, which is incorporated by reference herein. Additionally, feature data can also include texture class data that allows users to perceive contrast between images, such as edges or other areas of high contrast. Although the present invention will be described with regard to the CIE Lab color model and a single texture feature, one skilled in the relevant art will appreciate that the disclosed embodiments are illustrative in nature and should not be construed as limiting.
  • Similar to the RGB color model and as described above, pixel display attributes may be expressed in a CIE Lab color model according to “L”, “a”, and “b” values. In the event that there is overlapping content between two or more images, in an illustrative embodiment of the present invention, the overlapping content can be considered a weighted combination of the individual “L”, “a”, “b” color values for the two or more overlapping images. Additionally, in an illustrative embodiment of the present invention, the overlapping content can also be represented according to a weighted combination of the texture values of both images. However, unlike traditional alpha blending, the weights applied to each feature may be different. In accordance with the present invention, Equation 2 defines the combination of “L”, “a”, and “b” values as:
    LxAxBx=(ω1L1+(1−ω1)L2)+(ω2a1+(1−ω2)a2)+(ω3b1+(1−ω3)b2)  (2)
    where:
      • L1a1b1=CIE Lab color pixel values for a first image;
      • L2a2b2=CIE Lab color pixel values for a second image;
      • ω1=weight value for L channel;
      • ω2=weight value for a channel; and
      • ω3=weight value for b channel.
  • In an illustrative embodiment of the present invention, ω1, ω2, and ω3, are selected in a manner that facilitates user recognition of the perception of each image. The individual weight values may correspond to a range of numbers from 0 to 1. Additionally, in an actual embodiment of the present invention, each weight value corresponds to a Boolean value of either “1” or “0”. By limiting weight values to Boolean values, each feature class will correspond only to the channel values from one of the overlapping images. In an illustrative embodiment of the present invention, Equation 2 may be modified to include texture class data by the inclusion of a weight value or each texture class channel. Equation 3 corresponds to the modification of Equation 2 to include one or more channels of texture class data.
    LxAxBx+τ1−tX—=(ω1L1+(1−ω1)L2)+(ω2a1+(1−ω2)a2)+(ω3b1+(1−ω3)b2)+(ω4τ11+(1−ω4)τ21) . . . +(ω 1 X+(1−ωX2 X)  (3)
    where:
      • L1a1b1=CIE Lab color pixel values for a first image;
      • L2a2b2=CIE Lab color pixel values for a second image;
      • τ11−τ1X=texture channel values for the first image;
      • τ21−τ2X=texture channel values for the second image
      • ω1=weight value for L channel;
      • ω2=weight value for a channel;
      • ω3=weight value for b channel;
      • ω4=weight value for first texture channel; and
      • ωX=weight value for Xth texture channel.
        Similar to the color channel data, the weight values for the texture classes can be expressed as Boolean value such that only one of the images can correspond to each texture feature class.
  • In an illustrative embodiment of the present invention, the individual channel weights may be pre-assigned by a user, by the software application, or according to the content being rendered. For example, the individual channel weights may relate to the type of overlapping content (e.g., overlapping palette graphical windows) to achieve a specific effect. An example of the blending of overlapping palette graphical windows will be described in greater detail below. Additionally, the channel weight values may be expressed as a series of weight values that may be manually or dynamically modified. For example, a computer may utilize a series of channel weights to achieve a transition between blended images.
  • With reference now to FIG. 3, a routine 300 for blending multiple images will be described. At block 302, a computer system associates two or more images to be blended with a set of feature classes. As described above, in an illustrative embodiment of the present invention, each pixel in an image may be represented according to its L, a and b CIE Lab color model value. Additionally, each pixel in an image may also be represented by one or more texture values as well. In the event that the pixel information for an image is represented in another color model, such RGB, the routine 300 can include the translation of pixel color model data from another other color model to the CIE Lab color model.
  • At block 304, the computer system associates channel weights for each of the feature classes. As described above, in an illustrative embodiment of the present invention, the channel weights may be pre-assigned based on user input, computer settings and/or the type of windows being blended. For example, a graphical window corresponding to a palette may be blended with different channel weights than two overlapping graphical windows corresponding to instantiated software applications. As also described above, in an illustrative embodiment of the present invention, the individual channel weights can be expressed as Boolean values.
  • At block 306, the two images are processed according to the particular channel weights to generate a combined set of feature values for each pixel. In an illustrative embodiment of the present invention, the image pixel data may be processed in a manner corresponding to Equation 2. For example, if the channel weights ω1, ω2, ω3, are set to “1”, “0” and “1” values, the resulting pixel image would utilize the L and b values from the first image and the “a” color value from the second image. Additionally, if the texture channel weight is set to “0”, the texture values, e.g., the edge values from the second image will be used. At block 308, the net set of feature values, e.g., color values and the texture values, may be adjusted to improve image displays. One skilled in the relevant art will appreciate that a number of image processing techniques/processes may be used to improve overall image displays. In one embodiment, image processing filters, such as contrast filters, color filters, blur filters and the like may be utilized to adjust the foreground or background images. In another embodiment, and as will be explained in greater detail below, channel data from one of the images may be ported to another channel to distinguish the foreground and background images. In a further embodiment, usage data may be utilized to diminish portions of a foreground image in favor of underlying content associated with a background image. In still a further embodiment, other traditional image processing techniques, such as alpha blending may also be utilized to improve the image displays. At block 310, the routine 300 ends.
  • In an illustrative embodiment of the present invention, the multi-blending techniques described above may be utilized for representing graphical windows corresponding to a palette. FIG. 4 is a flow diagram illustrative of a palette multi-blending routine 400 implemented by a computer system in accordance with the present invention. Routine 400 corresponds to block 308 in which color channel weight values have been set for two images in which the background image corresponds to a photograph or other content and the foreground image corresponds to a palette. Based on the characteristics of the content, channel weights ω1, ω2, ω3, are set to “1”, “0” and “0”. Additionally, the texture values are set to “1.”
  • With reference to FIG. 4, at block 402, the images are desaturated to apply the color channel weights. With reference to the working example, because of the channel weight settings, the blending image includes the red/green and blue/yellow values from the background image. However, the luminance values correspond solely to the foreground image. Thus, the background image colors are preserved and not diluted. At block 404, the palette surfaces are made transparent to apply the texture channel settings. In an illustrative embodiment of the present invention, palettes typically consist of icons corresponding to controls, whose contours can be recognized by users. Accordingly, to apply the texture channel settings, the image is filtered to with a high-pass filter, such as an emboss filter, to bring out the edges of the controls. In essence, the image is filtered to call out its texture channel values. In an illustrative embodiment of the present invention, additional blending techniques, such as the linear light blending technique, may be applied to preserve light/dark contrast.
  • As described above, blocks 402 and 404 correspond to the application of Boolean weight channels. With continued reference to FIG. 4, in an illustrative embodiment of the present invention, the routine 400 can include additional processes that improve the overall improvement of the blended image (block 308, FIG. 3). One skilled in the relevant art will appreciate that the image improvement processes may be optional and/or dependent on the type of image content being blended. For example, the blending of two graphic images may require a different set of improvement processes than the blending of two textual messages. At block 406, a portion of the background image is blurred to eliminate high frequency interference associated with noisy backgrounds. More specifically, the overlapping portion of the background image may be blurred with a blur filter. Further, in an illustrative embodiment of the present invention, the blur filter may be configured to dynamically blur any portion of the background image overlapping with the foreground image as the foreground image moves. Thus, the resulting image provides high contrast for the foreground image and low contrast for the overlapping portion of the background image.
  • At block 408, usage data may be utilized to vary opacity of portions of the foreground image. In an illustrative embodiment of the present invention, usage data corresponding to selection data or other explicit or implicit user feedback systems, is used to vary portions of the display. For example, borders, unused icons, or persistent icons that a user is familiar with from a foreground image, such as a palette, may be diminished in favor of the background image.
  • At block 410, one or more feature channels may be remapped in the event of conflicting images. In an illustrative embodiment of the present invention, the foreground and background image may correspond to two images utilizing similar features to represent the content. For example, two images corresponding primarily to text may rely heavily on the luminance channel to represent. As described above, however, in an illustrative embodiment of the present invention, only one of the images will be able to use its luminance values because they otherwise interfere. Accordingly, one of the image's luminance values may be mapped to another channel, such as the red/green and blue/yellow difference values to mitigate interference. One skilled in the relevant art will appreciate that additional adjustment processes may also be utilized. At block 412, the routine 400 terminates.
  • FIG. 5 is a block diagram illustrative of a portion of a screen display 500 illustrating the results of processing a foreground 502 and background image 504 in accordance with the present invention. As illustrated in FIG. 5, user recognition of the background image 504 is enhanced by preserving color channel values. Additionally, foreground image 502 recognition is achieved by preserving luminance and texture channel values. One skilled in the relevant art will appreciate the resulting screen display 500 is illustrative in nature and that alternative results may also be achieved within the scope of the present invention.
  • While illustrative embodiments of the invention have been illustrated and described, it will be appreciated that various changes can be made therein without departing from the spirit and scope of the invention.

Claims (32)

1. A method for processing two or more images represented by a set of feature classes, the method comprising:
associating each of the two or more images with a set of feature class data;
assigning a processing weight for each feature class; and
processing the set of feature class data according to the assigned processing weight to generate a blended image.
2. The method as recited in claim 1, wherein the set of feature class data corresponds to color class data.
3. The method as recited in claim 2, wherein the color class data corresponds to CIE Lab color model data.
4. The method as recited in claim 1, wherein the set of feature class data corresponds to texture class data.
5. The method as recited in claims 4, wherein the texture class data corresponds to edge data.
6. The method as recited in claim 1, wherein the processing weight for each feature class is specified as a range of numbers between zero and one.
7. The method as recited in claim 1, wherein the processing weight for each feature class is specified as a Boolean value.
8. A computer-readable medium having computer-executable instructions for performing the method recited in claim 1.
9. A computer system for performing the method recited in claim 1.
10. A method for blending a foreground image and a background image, wherein the foreground and background images are represented by a set of feature classes, the method comprising:
assigning a processing weight for each feature class;
processing the set of feature class data for the foreground and background image according to the assigned processing weight to generate a blended image; and
adjusting the blended image according to the content of the foreground and background image.
11. The method as recited in claim 10, wherein the set of feature class data corresponds to color class data.
12. The method as recited in claim 11, wherein the color class data corresponds to CIE Lab color model data.
13. The method as recited in claim 10, wherein the set of feature class data corresponds to texture class data.
14. The method as recited in claim 10, wherein the processing weight for each feature class is specified as a range of numbers between zero and one.
15. The method as recited in claim 10, wherein the processing weight for each feature class is specified as a Boolean value.
16. The method as recited in claim 10, wherein adjusting the blended image includes blurring the background image.
17. The method as recited in claim 10, wherein adjusting the blended image includes;
obtaining usage data for at least a portion of the foreground image; and
adjusting the image according to the usage data.
18. The method as recited in claim 17, wherein adjusting the image includes diminishing the opacity of portions of the foreground imaged based on the usage data.
19. The method as recited in claim 10, wherein adjusting the image includes mapping values from a first feature class of the foreground image to another feature class.
20. A computer-readable medium having computer-executable instructions for performing the method recited in claim 10.
21. A computer system for performing the method recited in claim 10.
22. A method for processing two or more images represented by a set of feature classes, the method comprising:
associating each of the two or more images with a set of feature class data;
assigning a processing weight for each feature class;
processing the set of feature class data according to the assigned processing weight to generate a blended image; and
adjusting the blended image according to the content of the two or more images.
23. The method as recited in claim 22, wherein the set of feature class data corresponds to color class data.
24. The method as recited in claim 23, wherein the color class data corresponds to CIE Lab color model data.
25. The method as recited in claim 22, wherein the processing weight for each feature class is specified as a range of numbers between zero and one.
26. The method as recited in claim 22, wherein the processing weight for each feature class is specified as a Boolean value.
27. The method as recited in claim 22, wherein adjusting the blended image includes blurring at least one of the two or more images.
28. The method as recited in claim 22, wherein adjusting the blended image includes;
obtaining usage data for at least a portion of one of the two or more images; and
adjusting the image according to the usage data.
29. The method as recited in claim 28, wherein adjusting the image includes diminishing the opacity of portions of the one of the two or more images based on the usage data.
30. The method as recited in claim 22, wherein adjusting the image includes mapping values from a first feature class of the foreground image to another feature class.
31. A computer-readable medium having computer-executable instructions for performing the method recited in claim 22.
32. A computer system for performing the method recited in claim 22.
US10/859,747 2003-06-12 2004-06-03 System and method for displaying images utilizing multi-blending Expired - Fee Related US7667717B2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
KR1020040040296A KR101076900B1 (en) 2003-06-12 2004-06-03 System and method for displaying images utilizing multi-blending
US10/859,747 US7667717B2 (en) 2003-06-12 2004-06-03 System and method for displaying images utilizing multi-blending

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US47829403P 2003-06-12 2003-06-12
US10/859,747 US7667717B2 (en) 2003-06-12 2004-06-03 System and method for displaying images utilizing multi-blending

Publications (2)

Publication Number Publication Date
US20050093890A1 true US20050093890A1 (en) 2005-05-05
US7667717B2 US7667717B2 (en) 2010-02-23

Family

ID=33418472

Family Applications (1)

Application Number Title Priority Date Filing Date
US10/859,747 Expired - Fee Related US7667717B2 (en) 2003-06-12 2004-06-03 System and method for displaying images utilizing multi-blending

Country Status (5)

Country Link
US (1) US7667717B2 (en)
EP (1) EP1489591B1 (en)
JP (1) JP4554280B2 (en)
KR (1) KR101076900B1 (en)
CN (1) CN100390825C (en)

Cited By (60)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060059432A1 (en) * 2004-09-15 2006-03-16 Matthew Bells User interface having viewing area with non-transparent and semi-transparent regions
US20070040849A1 (en) * 2005-08-19 2007-02-22 Eric Jeffrey Making an overlay image edge artifact less conspicuous
US20070065012A1 (en) * 2005-09-16 2007-03-22 Seiko Epson Corporation Image processing apparatus, image processing method, and program product
US20070296736A1 (en) * 2006-06-26 2007-12-27 Agfa Inc. System and method for scaling overlay images
US20080305795A1 (en) * 2007-06-08 2008-12-11 Tomoki Murakami Information provision system
US20090138811A1 (en) * 2005-11-02 2009-05-28 Masaki Horiuchi Display object penetrating apparatus
US7667717B2 (en) * 2003-06-12 2010-02-23 Microsoft Corporation System and method for displaying images utilizing multi-blending
US20110025701A1 (en) * 2009-07-23 2011-02-03 Samsung Electronics Co., Ltd. Method and system for creating an image
US8149247B1 (en) * 2007-11-06 2012-04-03 Nvidia Corporation Method and system for blending rendered images from multiple applications
US20130038744A1 (en) * 2011-08-09 2013-02-14 Adrian Proca Fast zero recovery with alpha blending in gyroscopic image stabilization
US20130083979A1 (en) * 2011-09-30 2013-04-04 Stefan Vilsmeier Method and device for displaying changes in medical image data
US9058653B1 (en) 2011-06-10 2015-06-16 Flir Systems, Inc. Alignment of visible light sources based on thermal images
US9143703B2 (en) 2011-06-10 2015-09-22 Flir Systems, Inc. Infrared camera calibration techniques
US9207708B2 (en) 2010-04-23 2015-12-08 Flir Systems, Inc. Abnormal clock rate detection in imaging sensor arrays
US9208542B2 (en) 2009-03-02 2015-12-08 Flir Systems, Inc. Pixel-wise noise reduction in thermal images
US9235023B2 (en) 2011-06-10 2016-01-12 Flir Systems, Inc. Variable lens sleeve spacer
US9235876B2 (en) 2009-03-02 2016-01-12 Flir Systems, Inc. Row and column noise reduction in thermal images
US9292909B2 (en) 2009-06-03 2016-03-22 Flir Systems, Inc. Selective image correction for infrared imaging devices
USD765081S1 (en) 2012-05-25 2016-08-30 Flir Systems, Inc. Mobile communications device attachment with camera
US9451183B2 (en) 2009-03-02 2016-09-20 Flir Systems, Inc. Time spaced infrared image enhancement
TWI553538B (en) * 2009-11-25 2016-10-11 雅虎股份有限公司 Gallery application for content viewing
CN106027888A (en) * 2016-05-20 2016-10-12 广东欧珀移动通信有限公司 Camera preview method for intelligent terminal and intelligent terminal
US9473681B2 (en) 2011-06-10 2016-10-18 Flir Systems, Inc. Infrared camera system housing with metalized surface
US9509924B2 (en) 2011-06-10 2016-11-29 Flir Systems, Inc. Wearable apparatus with integrated infrared imaging module
US9517679B2 (en) 2009-03-02 2016-12-13 Flir Systems, Inc. Systems and methods for monitoring vehicle occupants
US9521289B2 (en) 2011-06-10 2016-12-13 Flir Systems, Inc. Line based image processing and flexible memory system
US9635285B2 (en) 2009-03-02 2017-04-25 Flir Systems, Inc. Infrared imaging enhancement with fusion
US9674458B2 (en) 2009-06-03 2017-06-06 Flir Systems, Inc. Smart surveillance camera systems and methods
US9706139B2 (en) 2011-06-10 2017-07-11 Flir Systems, Inc. Low power and small form factor infrared imaging
US9706138B2 (en) 2010-04-23 2017-07-11 Flir Systems, Inc. Hybrid infrared sensor array having heterogeneous infrared sensors
US9706137B2 (en) 2011-06-10 2017-07-11 Flir Systems, Inc. Electrical cabinet infrared monitor
US9716843B2 (en) 2009-06-03 2017-07-25 Flir Systems, Inc. Measurement device for electrical installations and related methods
US9723227B2 (en) 2011-06-10 2017-08-01 Flir Systems, Inc. Non-uniformity correction techniques for infrared imaging devices
US9756264B2 (en) 2009-03-02 2017-09-05 Flir Systems, Inc. Anomalous pixel detection
US9756262B2 (en) 2009-06-03 2017-09-05 Flir Systems, Inc. Systems and methods for monitoring power systems
US9807319B2 (en) 2009-06-03 2017-10-31 Flir Systems, Inc. Wearable imaging devices, systems, and methods
WO2017189039A1 (en) * 2016-04-25 2017-11-02 Beach Lewis Image processing device and method
US9811884B2 (en) 2012-07-16 2017-11-07 Flir Systems, Inc. Methods and systems for suppressing atmospheric turbulence in images
US9819880B2 (en) 2009-06-03 2017-11-14 Flir Systems, Inc. Systems and methods of suppressing sky regions in images
US9843742B2 (en) 2009-03-02 2017-12-12 Flir Systems, Inc. Thermal image frame capture using de-aligned sensor array
US9848134B2 (en) 2010-04-23 2017-12-19 Flir Systems, Inc. Infrared imager with integrated metal layers
US9900526B2 (en) 2011-06-10 2018-02-20 Flir Systems, Inc. Techniques to compensate for calibration drifts in infrared imaging devices
US9948872B2 (en) 2009-03-02 2018-04-17 Flir Systems, Inc. Monitor and control systems and methods for occupant safety and energy efficiency of structures
US9961277B2 (en) 2011-06-10 2018-05-01 Flir Systems, Inc. Infrared focal plane array heat spreaders
US9973692B2 (en) 2013-10-03 2018-05-15 Flir Systems, Inc. Situational awareness by compressed display of panoramic views
US9986175B2 (en) 2009-03-02 2018-05-29 Flir Systems, Inc. Device attachment with infrared imaging sensor
US9998697B2 (en) 2009-03-02 2018-06-12 Flir Systems, Inc. Systems and methods for monitoring vehicle occupants
US10019737B2 (en) 2015-04-06 2018-07-10 Lewis Beach Image processing device and method
US10051210B2 (en) 2011-06-10 2018-08-14 Flir Systems, Inc. Infrared detector array with selectable pixel binning systems and methods
US10079982B2 (en) 2011-06-10 2018-09-18 Flir Systems, Inc. Determination of an absolute radiometric value using blocked infrared sensors
US10091439B2 (en) 2009-06-03 2018-10-02 Flir Systems, Inc. Imager with array of multiple infrared imaging modules
US10169666B2 (en) 2011-06-10 2019-01-01 Flir Systems, Inc. Image-assisted remote control vehicle systems and methods
US10244190B2 (en) 2009-03-02 2019-03-26 Flir Systems, Inc. Compact multi-spectrum imaging with fusion
US10389953B2 (en) 2011-06-10 2019-08-20 Flir Systems, Inc. Infrared imaging device having a shutter
US10757308B2 (en) 2009-03-02 2020-08-25 Flir Systems, Inc. Techniques for device attachment with dual band imaging sensor
US10841508B2 (en) 2011-06-10 2020-11-17 Flir Systems, Inc. Electrical cabinet infrared monitor systems and methods
US11094139B2 (en) * 2018-06-26 2021-08-17 Quantificare S.A. Method and device to simulate, visualize and compare surface models
US11297264B2 (en) 2014-01-05 2022-04-05 Teledyne Fur, Llc Device attachment with dual band imaging sensor
US20220300350A1 (en) * 2021-03-04 2022-09-22 Canon Kabushiki Kaisha Information processing apparatus, control method of information processing apparatus, and recording medium
US11568587B2 (en) * 2021-03-30 2023-01-31 International Business Machines Corporation Personalized multimedia filter

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1923838A1 (en) * 2006-11-20 2008-05-21 Agfa HealthCare NV Method of fusing digital images
JP4758950B2 (en) * 2007-06-07 2011-08-31 株式会社日立製作所 PLANT MONITORING DEVICE AND PLANT OPERATION MONITORING METHOD
US8806231B2 (en) * 2009-12-22 2014-08-12 Intel Corporation Operating system independent network event handling
US9524573B2 (en) * 2011-06-05 2016-12-20 Apple Inc. Systems, methods, and computer-readable media for manipulating and mapping tiles of graphical object data
CN102663806B (en) * 2012-03-02 2014-12-10 西安交通大学 Artistic-vision-based cartoon stylized rendering method of image
KR102285188B1 (en) * 2015-02-13 2021-08-05 삼성전자주식회사 Image processing method for improving the visibility and Electronic Device supporting the Same
CN105828156A (en) * 2016-03-22 2016-08-03 乐视网信息技术(北京)股份有限公司 Method and device of generating title background in video picture

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5517437A (en) * 1993-06-22 1996-05-14 Matsushita Electric Industrial Co., Ltd. Alpha blending calculator
US6380945B1 (en) * 1998-11-09 2002-04-30 Broadcom Corporation Graphics display system with color look-up table loading mechanism
US6384840B1 (en) * 1992-12-15 2002-05-07 Sun Microsystems, Inc. Method and apparatus for presenting information in a display system using transparent windows
US6518970B1 (en) * 2000-04-20 2003-02-11 Ati International Srl Graphics processing device with integrated programmable synchronization signal generation
US6587111B2 (en) * 1997-12-22 2003-07-01 Hitachi, Ltd. Graphic processor and data processing system
US20030137522A1 (en) * 2001-05-02 2003-07-24 Kaasila Sampo J. Innovations for the display of web pages
US20040042662A1 (en) * 1999-04-26 2004-03-04 Wilensky Gregg D. Identifying intrinsic pixel colors in a region of uncertain pixels
US6781591B2 (en) * 2001-08-15 2004-08-24 Mitsubishi Electric Research Laboratories, Inc. Blending multiple images using local and global information
US20050001852A1 (en) * 2003-07-03 2005-01-06 Dengler John D. System and method for inserting content into an image sequence
US20050088464A1 (en) * 2003-10-24 2005-04-28 Microsoft Corporation Fast rendering of ink
US6927778B2 (en) * 2002-05-16 2005-08-09 Ati Technologies, Inc. System for alpha blending and method thereof
US20050253877A1 (en) * 2004-05-12 2005-11-17 Thompson Robert D Display resolution systems and methods
US20050268226A1 (en) * 2004-05-28 2005-12-01 Lipsky Scott E Method and system for displaying image information

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2000028467A1 (en) * 1998-11-06 2000-05-18 The Trustees Of Columbia University In The City Of New York Image description system and method
CA2256970A1 (en) * 1998-12-23 2000-06-23 John-Paul J. Gignac Method for accessing and rendering an image
JP3441001B2 (en) * 2001-05-21 2003-08-25 富士ゼロックス株式会社 Image processing device
KR101076900B1 (en) * 2003-06-12 2011-10-25 마이크로소프트 코포레이션 System and method for displaying images utilizing multi-blending

Patent Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6384840B1 (en) * 1992-12-15 2002-05-07 Sun Microsystems, Inc. Method and apparatus for presenting information in a display system using transparent windows
US20020171682A1 (en) * 1992-12-15 2002-11-21 Sun Microsystems, Inc. Method and apparatus for presenting information in a display system using transparent windows
US5517437A (en) * 1993-06-22 1996-05-14 Matsushita Electric Industrial Co., Ltd. Alpha blending calculator
US6587111B2 (en) * 1997-12-22 2003-07-01 Hitachi, Ltd. Graphic processor and data processing system
US6380945B1 (en) * 1998-11-09 2002-04-30 Broadcom Corporation Graphics display system with color look-up table loading mechanism
US7002602B2 (en) * 1998-11-09 2006-02-21 Broadcom Corporation Apparatus and method for blending graphics and video surfaces
US6700588B1 (en) * 1998-11-09 2004-03-02 Broadcom Corporation Apparatus and method for blending graphics and video surfaces
US20040042662A1 (en) * 1999-04-26 2004-03-04 Wilensky Gregg D. Identifying intrinsic pixel colors in a region of uncertain pixels
US6518970B1 (en) * 2000-04-20 2003-02-11 Ati International Srl Graphics processing device with integrated programmable synchronization signal generation
US20030137522A1 (en) * 2001-05-02 2003-07-24 Kaasila Sampo J. Innovations for the display of web pages
US6781591B2 (en) * 2001-08-15 2004-08-24 Mitsubishi Electric Research Laboratories, Inc. Blending multiple images using local and global information
US6927778B2 (en) * 2002-05-16 2005-08-09 Ati Technologies, Inc. System for alpha blending and method thereof
US20050001852A1 (en) * 2003-07-03 2005-01-06 Dengler John D. System and method for inserting content into an image sequence
US20050088464A1 (en) * 2003-10-24 2005-04-28 Microsoft Corporation Fast rendering of ink
US20050253877A1 (en) * 2004-05-12 2005-11-17 Thompson Robert D Display resolution systems and methods
US20050268226A1 (en) * 2004-05-28 2005-12-01 Lipsky Scott E Method and system for displaying image information

Cited By (75)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7667717B2 (en) * 2003-06-12 2010-02-23 Microsoft Corporation System and method for displaying images utilizing multi-blending
US20060059432A1 (en) * 2004-09-15 2006-03-16 Matthew Bells User interface having viewing area with non-transparent and semi-transparent regions
US20070040849A1 (en) * 2005-08-19 2007-02-22 Eric Jeffrey Making an overlay image edge artifact less conspicuous
US7417647B2 (en) * 2005-08-19 2008-08-26 Seiko Epson Corporation Making an overlay image edge artifact less conspicuous
US20070065012A1 (en) * 2005-09-16 2007-03-22 Seiko Epson Corporation Image processing apparatus, image processing method, and program product
US7526139B2 (en) * 2005-09-16 2009-04-28 Seiko Epson Corporation Image processing for improving character readability of characters disposed on an image
US20090138811A1 (en) * 2005-11-02 2009-05-28 Masaki Horiuchi Display object penetrating apparatus
US20070296736A1 (en) * 2006-06-26 2007-12-27 Agfa Inc. System and method for scaling overlay images
US8072472B2 (en) * 2006-06-26 2011-12-06 Agfa Healthcare Inc. System and method for scaling overlay images
US20080305795A1 (en) * 2007-06-08 2008-12-11 Tomoki Murakami Information provision system
US8149247B1 (en) * 2007-11-06 2012-04-03 Nvidia Corporation Method and system for blending rendered images from multiple applications
US9948872B2 (en) 2009-03-02 2018-04-17 Flir Systems, Inc. Monitor and control systems and methods for occupant safety and energy efficiency of structures
US9235876B2 (en) 2009-03-02 2016-01-12 Flir Systems, Inc. Row and column noise reduction in thermal images
US9517679B2 (en) 2009-03-02 2016-12-13 Flir Systems, Inc. Systems and methods for monitoring vehicle occupants
US10244190B2 (en) 2009-03-02 2019-03-26 Flir Systems, Inc. Compact multi-spectrum imaging with fusion
US9843742B2 (en) 2009-03-02 2017-12-12 Flir Systems, Inc. Thermal image frame capture using de-aligned sensor array
US9635285B2 (en) 2009-03-02 2017-04-25 Flir Systems, Inc. Infrared imaging enhancement with fusion
US10033944B2 (en) 2009-03-02 2018-07-24 Flir Systems, Inc. Time spaced infrared image enhancement
US9998697B2 (en) 2009-03-02 2018-06-12 Flir Systems, Inc. Systems and methods for monitoring vehicle occupants
US9756264B2 (en) 2009-03-02 2017-09-05 Flir Systems, Inc. Anomalous pixel detection
US9208542B2 (en) 2009-03-02 2015-12-08 Flir Systems, Inc. Pixel-wise noise reduction in thermal images
US9986175B2 (en) 2009-03-02 2018-05-29 Flir Systems, Inc. Device attachment with infrared imaging sensor
US10757308B2 (en) 2009-03-02 2020-08-25 Flir Systems, Inc. Techniques for device attachment with dual band imaging sensor
US9451183B2 (en) 2009-03-02 2016-09-20 Flir Systems, Inc. Time spaced infrared image enhancement
US9807319B2 (en) 2009-06-03 2017-10-31 Flir Systems, Inc. Wearable imaging devices, systems, and methods
US9292909B2 (en) 2009-06-03 2016-03-22 Flir Systems, Inc. Selective image correction for infrared imaging devices
US9674458B2 (en) 2009-06-03 2017-06-06 Flir Systems, Inc. Smart surveillance camera systems and methods
US10091439B2 (en) 2009-06-03 2018-10-02 Flir Systems, Inc. Imager with array of multiple infrared imaging modules
US9756262B2 (en) 2009-06-03 2017-09-05 Flir Systems, Inc. Systems and methods for monitoring power systems
US9843743B2 (en) 2009-06-03 2017-12-12 Flir Systems, Inc. Infant monitoring systems and methods using thermal imaging
US9716843B2 (en) 2009-06-03 2017-07-25 Flir Systems, Inc. Measurement device for electrical installations and related methods
US9819880B2 (en) 2009-06-03 2017-11-14 Flir Systems, Inc. Systems and methods of suppressing sky regions in images
US20110025701A1 (en) * 2009-07-23 2011-02-03 Samsung Electronics Co., Ltd. Method and system for creating an image
US8830251B2 (en) 2009-07-23 2014-09-09 Samsung Electronics Co., Ltd. Method and system for creating an image
TWI553538B (en) * 2009-11-25 2016-10-11 雅虎股份有限公司 Gallery application for content viewing
US10324976B2 (en) 2009-11-25 2019-06-18 Oath Inc. Gallery application for content viewing
US9848134B2 (en) 2010-04-23 2017-12-19 Flir Systems, Inc. Infrared imager with integrated metal layers
US9207708B2 (en) 2010-04-23 2015-12-08 Flir Systems, Inc. Abnormal clock rate detection in imaging sensor arrays
US9706138B2 (en) 2010-04-23 2017-07-11 Flir Systems, Inc. Hybrid infrared sensor array having heterogeneous infrared sensors
US9706139B2 (en) 2011-06-10 2017-07-11 Flir Systems, Inc. Low power and small form factor infrared imaging
US9143703B2 (en) 2011-06-10 2015-09-22 Flir Systems, Inc. Infrared camera calibration techniques
US9723227B2 (en) 2011-06-10 2017-08-01 Flir Systems, Inc. Non-uniformity correction techniques for infrared imaging devices
US9716844B2 (en) 2011-06-10 2017-07-25 Flir Systems, Inc. Low power and small form factor infrared imaging
US9706137B2 (en) 2011-06-10 2017-07-11 Flir Systems, Inc. Electrical cabinet infrared monitor
US9538038B2 (en) 2011-06-10 2017-01-03 Flir Systems, Inc. Flexible memory systems and methods
US10841508B2 (en) 2011-06-10 2020-11-17 Flir Systems, Inc. Electrical cabinet infrared monitor systems and methods
US10389953B2 (en) 2011-06-10 2019-08-20 Flir Systems, Inc. Infrared imaging device having a shutter
US9521289B2 (en) 2011-06-10 2016-12-13 Flir Systems, Inc. Line based image processing and flexible memory system
US9509924B2 (en) 2011-06-10 2016-11-29 Flir Systems, Inc. Wearable apparatus with integrated infrared imaging module
US9473681B2 (en) 2011-06-10 2016-10-18 Flir Systems, Inc. Infrared camera system housing with metalized surface
US10250822B2 (en) 2011-06-10 2019-04-02 Flir Systems, Inc. Wearable apparatus with integrated infrared imaging module
US9900526B2 (en) 2011-06-10 2018-02-20 Flir Systems, Inc. Techniques to compensate for calibration drifts in infrared imaging devices
US10230910B2 (en) 2011-06-10 2019-03-12 Flir Systems, Inc. Infrared camera system architectures
US9961277B2 (en) 2011-06-10 2018-05-01 Flir Systems, Inc. Infrared focal plane array heat spreaders
US10169666B2 (en) 2011-06-10 2019-01-01 Flir Systems, Inc. Image-assisted remote control vehicle systems and methods
US9235023B2 (en) 2011-06-10 2016-01-12 Flir Systems, Inc. Variable lens sleeve spacer
US9723228B2 (en) 2011-06-10 2017-08-01 Flir Systems, Inc. Infrared camera system architectures
US10079982B2 (en) 2011-06-10 2018-09-18 Flir Systems, Inc. Determination of an absolute radiometric value using blocked infrared sensors
US9058653B1 (en) 2011-06-10 2015-06-16 Flir Systems, Inc. Alignment of visible light sources based on thermal images
US10051210B2 (en) 2011-06-10 2018-08-14 Flir Systems, Inc. Infrared detector array with selectable pixel binning systems and methods
US8872927B2 (en) * 2011-08-09 2014-10-28 Cisco Technology, Inc. Fast zero recovery with alpha blending in gyroscopic image stabilization
US20130038744A1 (en) * 2011-08-09 2013-02-14 Adrian Proca Fast zero recovery with alpha blending in gyroscopic image stabilization
US8781188B2 (en) * 2011-09-30 2014-07-15 Brainlab Ag Method and device for displaying changes in medical image data
US20130083979A1 (en) * 2011-09-30 2013-04-04 Stefan Vilsmeier Method and device for displaying changes in medical image data
USD765081S1 (en) 2012-05-25 2016-08-30 Flir Systems, Inc. Mobile communications device attachment with camera
US9811884B2 (en) 2012-07-16 2017-11-07 Flir Systems, Inc. Methods and systems for suppressing atmospheric turbulence in images
US9973692B2 (en) 2013-10-03 2018-05-15 Flir Systems, Inc. Situational awareness by compressed display of panoramic views
US11297264B2 (en) 2014-01-05 2022-04-05 Teledyne Fur, Llc Device attachment with dual band imaging sensor
US10019737B2 (en) 2015-04-06 2018-07-10 Lewis Beach Image processing device and method
WO2017189039A1 (en) * 2016-04-25 2017-11-02 Beach Lewis Image processing device and method
CN106027888A (en) * 2016-05-20 2016-10-12 广东欧珀移动通信有限公司 Camera preview method for intelligent terminal and intelligent terminal
US11094139B2 (en) * 2018-06-26 2021-08-17 Quantificare S.A. Method and device to simulate, visualize and compare surface models
US20220300350A1 (en) * 2021-03-04 2022-09-22 Canon Kabushiki Kaisha Information processing apparatus, control method of information processing apparatus, and recording medium
US11822974B2 (en) * 2021-03-04 2023-11-21 Canon Kabushiki Kaisha Information processing apparatus, control method of information processing apparatus, and recording medium
US11568587B2 (en) * 2021-03-30 2023-01-31 International Business Machines Corporation Personalized multimedia filter

Also Published As

Publication number Publication date
JP4554280B2 (en) 2010-09-29
US7667717B2 (en) 2010-02-23
KR101076900B1 (en) 2011-10-25
KR20040104426A (en) 2004-12-10
EP1489591A3 (en) 2007-12-26
CN100390825C (en) 2008-05-28
JP2005135371A (en) 2005-05-26
EP1489591A2 (en) 2004-12-22
CN1704883A (en) 2005-12-07
EP1489591B1 (en) 2016-12-07

Similar Documents

Publication Publication Date Title
US7667717B2 (en) System and method for displaying images utilizing multi-blending
US6317128B1 (en) Graphical user interface with anti-interference outlines for enhanced variably-transparent applications
US7796139B1 (en) Methods and apparatus for displaying a frame with contrasting text
JP3615563B2 (en) How to draw scalable 3D boundaries
US7184063B2 (en) Adaptive color schemes
US6971071B1 (en) System and method for implementing an image ancillary to a cursor
US8791952B2 (en) Method and system of immersive generation for two-dimension still image and factor dominating method, image content analysis method and scaling parameter prediction method for generating immersive sensation
US7545389B2 (en) Encoding ClearType text for use on alpha blended textures
CN110473152A (en) Based on the image enchancing method for improving Retinex algorithm
USRE37476E1 (en) Alpha blending palettized image data
ES2616867T3 (en) Image display system and procedure using multiple mixing
US7969441B2 (en) Adaptive contextual filtering based on typographical characteristics
US7076112B2 (en) Image processing method
US20220036603A1 (en) Image processing for increasing visibility of obscured patterns
US11217205B2 (en) Method and apparatus for rendering contents for vision accessibility
Han et al. A novel confusion-line separation algorithm based on color segmentation for color vision deficiency
US8295539B2 (en) Method and system of immersive sensation enhancement for video sequence displaying
EP3836091A1 (en) Alpha value decision device, alpha value decision method, program, and data structure of image data
EP3503017A1 (en) Method and display device
US6295369B1 (en) Multi-dimensional color image mapping apparatus and method
JP2002351440A (en) Method and device for drawing gradational font
Hascoët Visual color design
Spelitz Color distribution transfer for mixed-reality applications
CN117032523A (en) Icon display method and device, electronic equipment and medium
Hong et al. Smart compositing: A real-time content-adaptive blending method for remote visual collaboration

Legal Events

Date Code Title Description
AS Assignment

Owner name: MICROSOFT CORPORATION, WASHINGTON

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:BAUDISCH, PATRICK M.;REEL/FRAME:015269/0693

Effective date: 20040901

Owner name: MICROSOFT CORPORATION,WASHINGTON

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:BAUDISCH, PATRICK M.;REEL/FRAME:015269/0693

Effective date: 20040901

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCF Information on status: patent grant

Free format text: PATENTED CASE

FPAY Fee payment

Year of fee payment: 4

AS Assignment

Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:034541/0477

Effective date: 20141014

FPAY Fee payment

Year of fee payment: 8

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20220223