US20050089218A1 - Digital camera having subject judgment function - Google Patents

Digital camera having subject judgment function Download PDF

Info

Publication number
US20050089218A1
US20050089218A1 US10/973,617 US97361704A US2005089218A1 US 20050089218 A1 US20050089218 A1 US 20050089218A1 US 97361704 A US97361704 A US 97361704A US 2005089218 A1 US2005089218 A1 US 2005089218A1
Authority
US
United States
Prior art keywords
image data
subject
reference image
dimensional distribution
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US10/973,617
Other versions
US7605847B2 (en
Inventor
Toru Chiba
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hoya Corp
Original Assignee
Pentax Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Pentax Corp filed Critical Pentax Corp
Assigned to PENTAX CORPORATION reassignment PENTAX CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CHIBA, TORU
Publication of US20050089218A1 publication Critical patent/US20050089218A1/en
Assigned to HOYA CORPORATION reassignment HOYA CORPORATION MERGER (SEE DOCUMENT FOR DETAILS). Assignors: PENTAX COPORATION
Application granted granted Critical
Publication of US7605847B2 publication Critical patent/US7605847B2/en
Assigned to Pentax Ricoh Imaging Company, Ltd. reassignment Pentax Ricoh Imaging Company, Ltd. CORPORATE SPLIT Assignors: HOYA CORPORATION
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/56Extraction of image or video features relating to colour

Definitions

  • FIG. 47 is a three-dimensional graph of the geometric averages in the eighth example.
  • FIG. 48 shows a subject image in a ninth example of the embodiment
  • the correction value setting program is a program for letting the CPU 70 a generate information for altering correction values regarding brightness, sharpness, color, etc. used by the image processor 30 into proper correction values corresponding to the subject information identified by the subject judgment program 73 and output the generated information to the image processor 30 .
  • step S 113 After obtaining the two-dimensional distribution data of the product-moment correlation coefficients regarding the Cr components of the subject image data and the Cr components of the reference image data as the processing object in the step S 113 , the CPU 70 a advances to step S 114 .
  • the CPU 70 a In the step S 114 , the CPU 70 a generates two-dimensional distribution data of product-moment correlation coefficients regarding the Cb components of the subject image data and the Cb components of the reference image data as the processing object.
  • the method for obtaining the two-dimensional distribution data in this step is similar to that in the step S 113 .
  • the CPU 70 a After obtaining the two-dimensional distribution data regarding the Cb components, the CPU 70 a advances to step S 115 .
  • the subject information is successively identified based on the image data successively supplied from the A/D converter 20 regardless of whether it is during shooting or not, by which the subject information stored in the prescribed area of the RAM 70 b is updated in succession.
  • FIG. 28 is a three-dimensional graph of the product-moment correlation coefficients regarding the Cr components in the fifth example
  • FIG. 29 is a three-dimensional graph of the product-moment correlation coefficients regarding the Cb components in the fifth example.
  • FIG. 30 is a table showing the geometric average of the product-moment correlation coefficients at each point in the two-dimensional coordinates
  • FIG. 31 is a three-dimensional graph of the geometric averages. Also in FIGS. 28 through 31 , negative correlation coefficients and geometric averages obtained therefrom are rounded to ⁇ 0.5. Also in the table of FIG. 30 , only 11 ⁇ 11 cells containing a characteristic part of the two-dimensional distribution are shown out of the two-dimensional distribution table having 41 ⁇ 41 cells.
  • FIG. 43 shows a subject image in an eighth example of the embodiment.
  • the subject matter of the subject image is the upper body of another person (different from that of the sixth example), and the reference image is the same as that of the sixth example (see FIG. 33 ).
  • the number of pixels of the subject image shown in FIG. 43 is 72 ⁇ 48.
  • the correspondence between the subject information i.e., a type of the subject
  • the parameters e.g., aperture value and shutter speed
  • the correspondence may be modified/input by a user of the digital still camera.

Abstract

A subject judgment program is capable of reducing the calculation load on a computer functioning as a device for judging the subject in a subject image. In a controller of a digital still camera, when a CPU running the subject judgment program (stored in a flash memory) obtains image data from an A/D converter via a first interface circuit, the CPU generates two-dimensional distribution data of product-moment correlation coefficients for each of model image data stored in the flash memory by successively calculating the product-moment correlation coefficient between the model image data and each part of the obtained image data, identifies a piece of model image data corresponding to two-dimensional distribution data having the highest maximum value, and thereby identifies subject information which has been associated with the identified model image data.

Description

    BACKGROUND OF THE INVENTION
  • The present invention relates to a subject judgment program for letting a computer, microprocessor, etc. function as a judging device that judges the type of a subject in an image formed by image data. The invention also relates to a digital camera which employs the subject judgment function to control an image capturing process and/or a quality correction process for correcting image data obtained or to be obtained by shooting a subject in accordance with the type of the subject.
  • As is commonly known, image data for displaying an image of a subject is generated when the subject is shot by a digital camera or when a positive/negative image containing the subject is read by a scanner, and the brightness, sharpness, color, etc. of the image formed by the image data can be changed by executing mathematical operations to the image data. Therefore, even when the quality of the image is not fine due to the settings/characteristics of the digital camera/scanner or the lighting on the subject, the image quality can be corrected easily only by executing proper computation to the image data.
  • Image processing software installed in digital cameras in recent years generally has such an image quality correction function. However, users having little experience in image quality correction do not understand which type of image quality correction process should be applied to the image data depending on the type of the subject in the image. Therefore, some of the many types of image processing software have the function of judging the type of the subject in the image formed by the image data in order to automatically execute the image quality correction during shooting or to inform the operator (user) of the type of the subject so that the operator can confirm the judgment is made correctly. An example of such process is described in Japanese Patent Provisional Publication No. 2003-018453.
  • However, the conventional image processing software described above is configured to judge the type of the subject based on setting/operational information obtained from each device in the digital camera such as a distance to the subject and a focal length and magnification of the lens used for shooting. Such a judgment may not be reliably correct since various subjects can be photographed at various operational settings.
  • It may be possible to introduce a so-called “matching technique” as a method for correctly judging the subject in the image. In this case, a lot of pieces of image data (images) of a variety of subjects are prepared as reference image data (reference images), and when image data of a subject image (as the target of image quality correction) is input, characteristic values of the subject image are compared with those of each reference image to obtain a difference regarding the characteristic values, and one of the reference images giving the smallest difference is identified. By such a process, the subject in the subject image is judged to be identical with the subject in the identified reference image.
  • In such a conventional method, in order to make the judgment as correct as possible, a large number of pieces of reference image data (reference images) may be prepared, and comparison of the input image and each of the reference images should be performed precisely.
  • However, digital cameras (especially, compact digital cameras) have certain limitations in storage capacity and processing speed for the image processing. Therefore, if the conventional matching process is performed, it takes a very long time for judgment, keeping the operator waiting for the completion of the judgment on the subject in the image. If the waiting time is shortened, the results may be unreliable.
  • SUMMARY OF THE INVENTION
  • The present invention is advantageous in that a subject judgment program is provided, which is capable of reducing the load on a computer, microprocessor, etc. functioning as the device for judging the type of the subject in the subject image, and a digital camera capable of controlling an image capturing process and/or the image quality correction process quickly with the precise judgment of the type of the subject.
  • According to an aspect of the invention, there is provided a computer program product comprising computer-readable instructions that cause a computer to function as a storage module which stores at least one piece of reference image data in a storage device associated with subject information which defines a subject in an image represented by the reference image data, an input module to which subject image data is input, a distribution data generation module which generates two-dimensional distribution data of product-moment correlation coefficients for at least one piece of reference image data stored in the storage device by calculating the product-moment correlation coefficients between the at least one reference image data and parts of the subject image data, an identification module which identifies a piece of two-dimensional distribution data having a value greater than a predetermined value, and an output module which reads out the subject information associated with the reference image data corresponding to the identified two-dimensional distribution data from the storage device and which outputs the subject information.
  • In accordance with another aspect of the present invention, there is provided a computer program product comprising computer-readable instructions that cause a computer to function as a storage module which stores one or more pieces of reference image data in a storage device associating each reference image data with subject information which defines a subject in an image formed by the reference image data, an input module to which subject image data is input, a distribution data generation module which generates two-dimensional distribution data of product-moment correlation coefficients for each of the one or more pieces of reference image data stored in the storage device by calculating the product-moment correlation coefficients between each reference image data and parts of the subject image data, an identification module which identifies a piece of two-dimensional distribution data having the highest maximum value among all the two-dimensional distribution data generated for all the reference image data, and an output module which reads out the subject information associated with the reference image data corresponding to the identified two-dimensional distribution data from the storage device and which outputs the subject information.
  • Preferably, the distribution data generation module includes a color space conversion module which converts color space of the subject image data into YCrCb while converting color space of all the reference image data stored in the storage device into YCrCb, a coefficient calculation module which calculates product-moment correlation coefficients between Cr components of each reference image data and parts of Cr components of the subject image data and thereby generates two-dimensional distribution data of the product-moment correlation coefficients regarding the Cr components for each reference image data, and which calculates product-moment correlation coefficients between Cb components of each reference image data and parts of Cb components of the subject image data and thereby generates two-dimensional distribution data of the product-moment correlation coefficients regarding the Cb components for each reference image data, and an average calculation module which executes an average calculation process, for calculating each geometric average of the product-moment correlation coefficients regarding the Cr and cb components at the same two-dimensional coordinates, for each reference image data and thereby generates two-dimensional distribution data of the geometric averages for each reference image data. In this case, the identification module identifies a piece of two-dimensional distribution data of geometric averages having the highest maximum value among all the two-dimensional distribution data of geometric averages generated for all the reference image data.
  • In accordance with another aspect of the present invention, there is provided a digital camera provided with a storage unit which stores one or more pieces of reference image data associating each reference image data with subject information which defines a subject in an image formed by the reference image data, an image pickup device which picks up a subject image formed by an objective optical system and thereby successively generates subject image data, an image processing unit which changes image quality of an image displayed on a display device according to the subject image data by executing computation to the subject image data using proper correction values, a distribution data generation unit which generates two-dimensional distribution data of product-moment correlation coefficients for each of the one or more pieces of reference image data stored in the storage unit by calculating the product-moment correlation coefficients between each reference image data and parts of the subject image data, an identification unit which identifies a piece of two-dimensional distribution data having the highest maximum value among all the two-dimensional distribution data generated for all the reference image data, a readout unit which reads out the subject information associated with the reference image data corresponding to the identified two-dimensional distribution data from the storage unit, and a setting alteration unit which alters the correction values used by the image processing unit into correction values corresponding to the subject information read out by the readout unit.
  • According to another aspect of the invention, there is provided a method of identifying a type of a subject image, which includes (a) preparing a plurality of pieces of reference image data respectively representing a plurality of different types of images, the number of pixels of the subject image data being different from the number of pixels of each of the reference image data, (b) scanning, of a piece of reference image data and subject image data, the data having the lower number of pixels within the other of the piece of reference image data and the subject image data to evaluate similarity therebetween at every predetermined scanning position to obtain two dimensional distribution data, the scanning being repeated for each of the plurality of pieces of reference image data to obtain the two dimensional distribution data for each of the plurality of pieces of the reference image data, and (c) identifying one of the plurality of pieces of the reference image data most similar to the subject image in accordance with the plurality of pieces of the two dimensional distribution data corresponding to the plurality of pieces of the reference image data.
  • In a particular case, the number of the pixels of each piece of the reference image data is smaller than the number of the pixels of the subject image data.
  • Optionally, the numbers of the pixels of the plurality of pieces of the reference image data may be the same.
  • Still optionally, the preparing may include generating each of the plurality of pieces of the reference image data by reducing the number of pixels of a plurality of pieces of predetermined model image data.
  • Further, the preparing may include reducing the number of pixels of an object image to generate the subject image.
  • Still optionally, the identifying may identify one of the plurality of pieces of the reference image data which corresponds to the two-dimensional distribution data having the highest maximum value of any two-dimensional distribution data as the reference image data corresponding to the subject image.
  • BRIEF DESCRIPTION OF THE ACCOMPANYING DRAWINGS
  • The file of this patent contains at least one drawing executed in color: Copies of this patent with color drawing(s) will be provided by the Patent and Trademark Office upon request and payment of the necessary fee.
  • FIG. 1 is a block diagram schematically showing the internal composition of a digital still camera in accordance with an embodiment of the present invention,
  • FIG. 2 is a block diagram schematically showing functional blocks of a controller of the digital still camera,
  • FIG. 3 is a data structure table for conceptually explaining a correspondence table of the digital still camera,
  • FIG. 4 is a flow chart explaining the contents of a subject judgment process,
  • FIG. 5 shows a subject image in a first example of the embodiment,
  • FIG. 6 shows a reference image in the first example,
  • FIG. 7 is a three-dimensional graph of product-moment correlation coefficients regarding Cr components in the first example,
  • FIG. 8 is a three-dimensional graph of product-moment correlation coefficients regarding Cb components in the first example,
  • FIG. 9 is a table showing a geometric average of the product-moment correlation coefficients at each point of two-dimensional coordinates in the first example,
  • FIG. 10 is a three-dimensional graph of the geometric averages in the first example,
  • FIG. 11 shows a subject image in a second example of the embodiment,
  • FIG. 12 is a three-dimensional graph of product-moment correlation coefficients regarding Cr components in the second example,
  • FIG. 13 is a three-dimensional graph of product-moment correlation coefficients regarding Cb components in the second example,
  • FIG. 14 is a table showing a geometric average of the product-moment correlation coefficients at each point of two-dimensional coordinates in the second example,
  • FIG. 15 is a three-dimensional graph of the geometric averages in the second example,
  • FIG. 16 shows a subject image in a third example of the embodiment,
  • FIG. 17 is a three-dimensional graph of product-moment correlation coefficients regarding Cr components in the third example,
  • FIG. 18 is a three-dimensional graph of product-moment correlation coefficients regarding Cb components in the third example,
  • FIG. 19 is a table showing a geometric average of the product-moment correlation coefficients at each point of two-dimensional coordinates in the third example,
  • FIG. 20 is a three-dimensional graph of the geometric averages in the third example,
  • FIG. 21 shows a subject image in a fourth example of the embodiment,
  • FIG. 22 is a three-dimensional graph of product-moment correlation coefficients regarding Cr components in the fourth example,
  • FIG. 23 is a three-dimensional graph of product-moment correlation coefficients regarding Cb components in the fourth example,
  • FIG. 24 is a table showing a geometric average of the product-moment correlation coefficients at each point of two-dimensional coordinates in the fourth example,
  • FIG. 25 is a three-dimensional graph of the geometric averages in the fourth example,
  • FIG. 26 shows a subject image in a fifth example of the embodiment,
  • FIG. 27 shows a reference image in the fifth example,
  • FIG. 28 is a three-dimensional graph of product-moment correlation coefficients regarding Cr components in the fifth example,
  • FIG. 29 is a three-dimensional graph of product-moment correlation coefficients regarding Cb components in the fifth example,
  • FIG. 30 is a table showing a geometric average of the product-moment correlation coefficients at each point of two-dimensional coordinates in the fifth example,
  • FIG. 31 is a three-dimensional graph of the geometric averages in the fifth example,
  • FIG. 32 shows a subject image in a sixth example of the embodiment,
  • FIG. 33 shows a reference image in the sixth example,
  • FIG. 34 is a three-dimensional graph of product-moment correlation coefficients regarding Cr components in the sixth example,
  • FIG. 35 is a three-dimensional graph of product-moment correlation coefficients regarding Cb components in the sixth example,
  • FIG. 36 is a table showing a geometric average of the product-moment correlation coefficients at each point of two-dimensional coordinates in the sixth example,
  • FIG. 37 is a three-dimensional graph of the geometric averages in the sixth example,
  • FIG. 38 shows a subject image in a seventh example of the embodiment,
  • FIG. 39 is a three-dimensional graph of product-moment correlation coefficients regarding Cr components in the seventh example,
  • FIG. 40 is a three-dimensional graph of product-moment correlation coefficients regarding Cb components in the seventh example,
  • FIG. 41 is a table showing a geometric average of the product-moment correlation coefficients at each point of two-dimensional coordinates in the seventh example,
  • FIG. 42 is a three-dimensional graph of the geometric averages in the seventh example,
  • FIG. 43 shows a subject image in an eighth example of the embodiment,
  • FIG. 44 is a three-dimensional graph of product-moment correlation coefficients regarding Cr components in the eighth example,
  • FIG. 45 is a three-dimensional graph of product-moment correlation coefficients regarding Cb components in the eighth example,
  • FIG. 46 is a table showing a geometric average of the product-moment correlation coefficients at each point of two-dimensional coordinates in the eighth example,
  • FIG. 47 is a three-dimensional graph of the geometric averages in the eighth example,
  • FIG. 48 shows a subject image in a ninth example of the embodiment,
  • FIG. 49 shows a reference image in the ninth example,
  • FIG. 50 is a three-dimensional graph of product-moment correlation coefficients regarding Cr components in the ninth example,
  • FIG. 51 is a three-dimensional graph of product-moment correlation coefficients regarding Cb components in the ninth example,
  • FIG. 52 is a table showing a geometric average of the product-moment correlation coefficients at each point of two-dimensional coordinates in the ninth example,
  • FIG. 53 is a three-dimensional graph of the geometric averages in the ninth example,
  • FIG. 54 is a flowchart illustrating an image capturing process which is executed when a shutter button of the camera is depressed, and
  • FIG. 55 is a flowchart illustrating an image compensating procedure called in the image capturing process shown in FIG. 54.
  • DESCRIPTION OF THE EMBODIMENT
  • Referring now to the drawings, a description will be given in detail of preferred embodiment in accordance with the present invention.
  • FIG. 1 is a block diagram schematically showing the internal composition of a digital still camera in accordance with an embodiment of the present invention. The digital still camera includes an image pickup device 10, an A/D (analog-to-digital) converter 20, an image processor 30, a storage 40, a display 50, a control panel 60 and a controller 70 as main components.
  • The image pickup device 10 is a device for picking up a subject image (an image of a subject) which is formed by an objective optical system 11. The image pickup device 10 constantly outputs image data obtained by the image pickup operation. The A/D converter 20 is a device for successively converting the image data of an analog signal format outputted by the image pickup device 10 into image data of a digital signal format.
  • The image processor 30 is a device for executing prescribed image processing to the image data outputted by the A/D converter 20. The image processor 30 obtains the image data from the A/D converter 20 in response to an instruction from the controller 70 (explained later) and changes the image quality of the image formed by the image data by applying proper mathematical operations corresponding to information supplied from the controller 70 to the obtained image data.
  • The storage 40 is a memory for successively storing the image data outputted by the image processor 30. The display 50 is a device for displaying an image based on the image data outputted by the image processor 30.
  • The control panel 60 is provided with various switches, buttons and keys for operating the digital still camera, including a switch for turning on/off the main power, switches for changing the shooting mode, display mode, etc., a shutter button, a cross-shaped key and buttons for altering various settings, etc.
  • The controller 70 controls the whole digital still camera executes various processes in response to inputs by the operator to the switches, buttons and keys on the control panel 60. FIG. 2 is a block diagram schematically showing functional blocks of the controller 70.
  • The controller 70 includes a CPU 70 a, a RAM 70 b, a first interface circuit 70 c, a second interface circuit 70 d, a third interface circuit 70 e and a flash memory 70 f. The hardware components 70 a-70 f are connected together by a bus B.
  • The CPU 70 a is a central processing unit for controlling the whole controller 70. The RAM 70 b is a main memory in which work areas are reserved when various programs are run by the CPU 70 a.
  • The first interface circuit 70 c is a device in charge of the reception of the image data from the A/D converter 20. The A/D converter 20 constantly outputs the image data. The image data received by the first interface circuit 70 c is successively transferred to the RAM 70 b.
  • The second interface circuit 70 d is a device in charge of the reception of information which is outputted by the control panel 60 when one or more of the switches, buttons and keys on the control panel 60 are operated.
  • The third interface circuit 70 e is a device in charge of the transmission of information to the image processor 30. The information transmitted by the third interface circuit 70 e to the image processor 30 is obtained by the CPU 70 a by running a program which will be explained later.
  • The flash memory 70 f is a storage device storing various programs and data. The flash memory 70 f is detachably attached to a socket of a circuit board which is a main component of the controller 70. Thus, the flash memory 70 f can be replaced with a new flash memory 70 f storing various programs and data which have been upgraded, updated or modified to new versions. In this case, the flash memory 70 f functions as an exchangeable and computer-readable record medium.
  • The data stored in the flash memory 70 f include one or more pieces of model image data 71. The model image data 71 are image data of images of subjects such as a human figure, red leaves of autumn, flowers, ocean, sunset and buildings.
  • The model image data 71 may be image data generated by the digital still camera by shooting subjects and then stored in the flash memory 70 f, image data input from another computer to the flash memory 70 f via an unshown communication function of the digital still camera, or image data stored in the flash memory 70 f when the digital still camera is manufactured and shipped from the factory.
  • The data stored in the flash memory 70 f also include a correspondence table 72. The correspondence table 72 is a table storing the correspondence between each model image data 71 and each subject information defining the subject in the model image formed by the model image data 71. FIG. 3 is a data structure table for conceptually explaining the correspondence table 72.
  • As shown in FIG. 3, the correspondence table 72 stores each record which is uniquely associated with each model image data 71. Each record includes a “file name” field and a “subject information” field. The “file name” field stores the file name of the model image data 71, while the “subject information” field stores the subject information regarding the type of the subject.
  • The programs stored in the flash memory 70 f include an overall control program for controlling the whole of the digital still camera. The overall control program includes a subject judgment program 73 in accordance with the present invention, a correction value setting program and a subject information display program, as its module programs.
  • The subject judgment program 73 is a program for letting the CPU 70 a judge the subject in each image formed by each image data successively supplied from the A/D converter 20 to the controller 70 and thereby identify the subject information on the subject.
  • The correction value setting program is a program for letting the CPU 70 a generate information for altering correction values regarding brightness, sharpness, color, etc. used by the image processor 30 into proper correction values corresponding to the subject information identified by the subject judgment program 73 and output the generated information to the image processor 30.
  • The subject information display program is a program for letting the CPU 70 a generate image data for superimposing the subject information (identified by the subject judgment program 73) on the image displayed on the display 50 and send the generated image data to the display 50.
  • In the following, a process carried out by the controller 70 composed as above will be described in detail.
  • When the main power of the digital still camera is turned on by operating the main power switch of the control panel 60, the CPU 70 a of the controller 70 reads the overall control program from the flash memory 70 f and thereby starts controlling the whole digital still camera. When the operation mode of the digital still camera is switched to a shooting mode by operating the mode switch of the control panel 60, the CPU 70 a reads the subject judgment program 73 from the flash memory 70 f, by which a subject judgment process is started by the controller 70. FIG. 4 is a flowchart explaining the contents of the subject judgment process.
  • In the first step S101 of the subject judgment process, the CPU 70 a waits until image data supplied from the A/D converter 20 is stored in the RAM 70 b via the first interface circuit 70 c. When image data of an image has been stored in the RAM 70 b, the CPU 70 a advances to step S102.
  • In the step S102, the CPU 70 a simplifies the image data stored in the R 70 b. According to the embodiment, the simplification executed in this step includes thinning out (omission) of data. It should be noted that other methods such as thumbnailing (scaling down), averaging, extraction of pixels satisfying specific conditions, etc. may be employed for simplification. However, the thinning out of data is advantageous since the simplified data retains the color of the original image.
  • By the simplification, the CPU 70 a reduces the number of pixels of the image data from several millions to 32×32 or 64×64, for example. Such image data with a reduced number of pixels will hereinafter be called “subject image data”. After generating the subject image data, the CPU 70 a advances to step S103.
  • In the step S103, the CPU 70 a converts the color space of the subject image data from RGB to YCrCb. Conceptually, the CPU 70 a generates color tone values (Y, Cr, Cb) of each pixel of a new subject image data by substituting color tone values (R, G, B) of the pixel of the (old) subject image data into the following three equations:
    Y=0.299R+0.587G+0.114B
    Cr=0.701R−0.587G−0.114B
    Cb=−0.299R−0.587G+0.886B
  • After the color space conversion of the step S103, the CPU 70 a executes a process loop L1 with respect to each model image data 71 stored in the flash memory 70 f.
  • In the process loop L1, the CPU 70 a designates each model image data 71 in the flash memory 70 f in turn as the object of the process and executes steps S111-S116 to each of the designated model image data 71 in sequence.
  • In the step S111, the CPU 70 a simplifies the model image data 71 as the processing object. The simplification executed in this step is the same as the simplification in the step S102. By such simplification, the CPU 70 a reduces the number of pixels of the model image data 71 to 10×10 or 20×20, for example. Such model image data with a reduced number of pixels will hereinafter be called “reference image data”. After generating the reference image data, the CPU 70 a advances to step S112.
  • In the step S112, the CPU 70 a converts the color space of the reference image data from RGB to YCrCb. This process is the same as the process executed in the step S103. After the color space conversion, the CPU 70 a advances to step S113.
  • In the step S113, the CPU 70 a generates two-dimensional distribution data of product-moment correlation coefficients regarding the Cr components of the subject image data and the Cr components of the reference image data as the processing object.
  • For example, when the subject image data has 20×20 pixels and the reference image data has 10×10 pixels, the CPU 70a first designates or extracts an area having 10×10 pixels from the subject image data and calculates a product-moment correlation coefficient between the 10×10 pixels of the designated area and the reference image data having 10×10 pixels. Subsequently, the CPU 70a repeats the process for calculating the product-moment correlation coefficient for all 10×10 pixel areas that can be extracted from the subject image data (allowing overlap between the areas). In this example, 11×11 product-moment correlation coefficients are obtained since 11×11 areas (having 10×10 pixels) can be extracted from the subject image data. Consequently, the CPU 70a generates two-dimensional distribution data having 11×11 product-moment correlation coefficients.
  • Incidentally, the “product-moment correlation coefficient” denotes a value for expressing the degree of correlation between a group X having n variables x and a group Y having n variables y. The product-moment correlation coefficient takes on values from −1 to 1. The method for calculating the product-moment correlation coefficient is well known in statistics. To describe the method briefly, the product-moment correlation coefficienty is obtained by dividing the covariance of the variables x and y by the product of the standard deviation of the variables x and the standard deviation of the variables y, as shown in the following expression. γ = 1 n - 1 i = 1 n ( Xi - X _ ) ( Yi - Y _ ) 1 n - 1 i = 1 n ( Xi - X _ ) 2 · 1 n - 1 i = 1 n ( Yi - Y _ ) 2
  • After obtaining the two-dimensional distribution data of the product-moment correlation coefficients regarding the Cr components of the subject image data and the Cr components of the reference image data as the processing object in the step S113, the CPU 70 a advances to step S114.
  • In the step S114, the CPU 70 a generates two-dimensional distribution data of product-moment correlation coefficients regarding the Cb components of the subject image data and the Cb components of the reference image data as the processing object. The method for obtaining the two-dimensional distribution data in this step is similar to that in the step S113. After obtaining the two-dimensional distribution data regarding the Cb components, the CPU 70 a advances to step S115.
  • In the step S115, the CPU 70 a calculates the geometric average of two corresponding (at the same two-dimensional coordinates) product-moment correlation coefficients in the two-dimensional distribution data regarding the Cr and Cb components (that is, a product-moment correlation coefficient C1 in the two-dimensional distribution data regarding the Cr components and a product-moment correlation coefficient C2 in the two-dimensional distribution data regarding the Cb components at the same two-dimensional coordinates as the product-moment correlation coefficient C1). The CPU 70 a generates two-dimensional distribution data of the geometric averages by executing the calculation of the geometric average for all the two-dimensional coordinates. For example, in the case where there are 11×11 product-moment correlation coefficients regarding the Cr components and 11×11 product-moment correlation coefficients regarding the Cb components, 11×11 geometric averages are calculated in the step S115. After obtaining the two-dimensional distribution data of the geometric averages, the CPU 70 a advances to step S116.
  • In the step S116, the CPU 70 a stores the two-dimensional distribution data of the geometric averages generated in the step S115 in the RAM 70 b, associating the two-dimensional distribution data with the subject information corresponding to the model image data 71 as the processing object, and ends the process loop L1 for the processing object.
  • The CPU 70 a repeats the above process loop L1 for all the model image data 71 existing in the flash memory 70 f and thereby generates the two-dimensional distribution data of the geometric averages for all the model image data 71. After exiting from the process loop L1, the CPU 70 a advances to step S104.
  • In the step S104, the CPU 70 a identifies (selects) a piece of two-dimensional distribution data of geometric averages having the highest maximum value of geometric average among all the two-dimensional distribution data of geometric averages generated in the step S115 during the process loops L1 for all the model image data 71. Thereafter, the CPU 70 a advances to step S105.
  • In the step S105, the CPU 70 a reads out subject information corresponding to the two-dimensional distribution data of geometric averages identified in the step S104 (i.e. subject information corresponding to the model image data 71 from which the two-dimensional distribution data of geometric averages identified in the step S104 was obtained) from the correspondence table 72. Thereafter, the CPU 70 a advances to step S106.
  • In the step S106, the CPU 70 a overwrites subject information stored in a prescribed area of the RAM 70 b with the subject information read out in the step S105 and thereby updates the subject information. After the update of the subject information, the CPU 70 a ends the subject judgment process.
  • By the above subject judgment process executed by the controller 70, the subject information is successively identified based on the image data successively supplied from the A/D converter 20 regardless of whether it is during shooting or not, by which the subject information stored in the prescribed area of the RAM 70 b is updated in succession.
  • When a shutter button of the control panel 60 is pressed under the above conditions, the correction value setting program is activated, by which the subject information existing in the prescribed area of the RAM 70 b at the point in time is read out, information for altering the correction values used by the image processor 30 into proper correction values corresponding to the subject information is generated, and the generated information is supplied to the image processor 30. The image processor 30 supplied with the information carries out an image quality correction process to the image data obtained from the A/D converter 20 during shooting, based on the correction values specified by the information.
  • In short, the subject is correctly identified and proper correction values corresponding to the type of the subject are set to the image processor 30 before the operator of the digital still camera presses the shutter button. Therefore, a proper image quality correction process corresponding to the type of the subject is executed quickly to the generated image data even when a variety of subjects are shot by the operator. Consequently, the operator is released from the need of waiting for a long time from the pressing of the shutter button (generation of the image data) to the storage of the image data in the storage 40.
  • Meanwhile, when the subject information stored in the prescribed area of the RAM 70 b is successively updated by the above subject judgment process, the CPU 70 a under the subject information display program generates information for superimposing the subject information on the image displayed on the display 50, by which the image with the subject information superposed thereon is displayed on the display 50.
  • Therefore, the operator can let the display 50 indicate the subject information corresponding to the subject only by aiming the digital still camera at the subject before pressing the shutter button on the control panel 60, by which the operator is allowed to easily grasp the type of the image quality correction process which is going to be executed to the subject to be shot.
  • Even when the correction values have to be set to the image processor 30 manually by the operator (with no automatic setting by the controller 70), the operator can see the subject information indicated on the display 50 and thereby learn which type of image quality correction should be done.
  • Incidentally, while the subject judgment program 73 is incorporated in the overall control program of the digital still camera in the above embodiment, the program in which the subject judgment program 73 can be incorporated is not limited to the overall control program of the digital still camera. For example, the subject judgment program 73 may be incorporated in image processing software that is installed in a personal computer. Also in such image processing software, by activating the subject judgment program 73, the subject information regarding a subject in an image of arbitrary image data can be presented to the user when the user is about to apply an image quality correction process to the image data. Therefore, the user can learn which type of image quality correction should be done based on the presented subject information.
  • In such cases, the subject judgment program 73 may be stored in a record medium (floppy disk, CD-ROM, DVD-ROM, etc.) and supplied to the personal computer to be read out and run by the personal computer. The subject judgment program 73 may also be supplied to the personal computer via a communication network such as the Internet. In the case where the subject judgment program 73 is installed in a digital still camera, the subject judgment program 73 may either be stored in the flash memory 70 f (record medium) when the digital still camera is manufactured and shipped from the factory or supplied to the digital still camera afterward via a record medium (Compact Flash, Smart Media, Memory Stick, etc.). The image data may also be input from an other computer to the flash memory 70 f via the unshown communication function of the digital still camera.
  • In the following, nine concrete examples of the two-dimensional distribution data of geometric averages, actually calculated and generated by the above subject judgment process, will be presented.
  • FIG. 5 shows a subject image in a first example of the embodiment, and FIG. 6 shows a reference image in the first example. In the first example shown in FIGS. 5 and 6, the subject matter of the subject image is the whole of a pale purple hydrangea at the tip of a spray, and the reference image is an image obtained by directly cutting out a part containing a few calyxes from the subject image. The number of pixels of the subject image shown in FIG. 5 is 64×64 and that of the reference image shown in FIG. 6 is 24×24. It should be noted that, in FIGS. 5 and 6 and other drawings, each image formed thereon consists of more pixels than the actual number of pixels for explanation purpose (i.e., in order to clearly show the object of each image).
  • FIG. 7 is a three-dimensional graph generated based on the two-dimensional distribution data of product-moment correlation coefficients regarding the Cr components of the subject image of FIG. 5 and the reference image of FIG. 6. FIG. 8 is a three-dimensional graph generated based on the two-dimensional distribution data of product-moment correlation coefficients regarding the Cb components of the subject image of FIG. 5 and the reference image of FIG. 6.
  • FIG. 9 is a table generated based on the two-dimensional distribution data regarding the Cr components and the two-dimensional distribution data regarding the Cb components, showing the geometric average of the product-moment correlation coefficients at each point in the two-dimensional coordinates. FIG. 10 is a three-dimensional graph generated based on the two-dimensional distribution data of the geometric averages.
  • Incidentally, in FIGS. 7 through 10, negative correlation coefficients and geometric averages obtained from the negative correlation coefficients are rounded to −0.5 for reducing the load of calculation and facilitating the drawing of the graphs. While the table showing the two-dimensional distribution of the geometric averages should have 41×41 cells statistically, only 11×11 cells containing a characteristic part of the two-dimensional distribution are shown in FIG. 9 due to limitations of space.
  • As shown in FIGS. 9 and 10, the maximum geometric average is 0.997 in the two-dimensional distribution of geometric averages in the first example. In other words, the two-dimensional distribution shows that the part of the subject image at the position with the maximum geometric average has an extremely high correlation with the reference image. Since the part of the subject image has exactly the same contents as the reference image, it can be said that the two-dimensional distribution of geometric averages is effectively serving as data properly representing the correlation between the subject image and the reference image.
  • FIG. 11 shows a subject image in a second example of the embodiment. In the second example, the subject matter of the subject image is a close-up of a pistil/stamen part of a red flower, and the reference image is the same as that of the first example (see FIG. 6). The number of pixels of the subject image shown in FIG. 11 is 64×64 similarly to the first example.
  • FIG. 12 is a three-dimensional graph of the product-moment correlation coefficients regarding the Cr components in the second example, and FIG. 13 is a three-dimensional graph of the product-moment correlation coefficients regarding the Cb components in the second example. FIG. 14 is a table showing the geometric average of the product-moment correlation coefficients at each point in the two-dimensional coordinates, and FIG. 15 is a three-dimensional graph of the geometric averages. Incidentally, also in FIGS. 12 through 15, negative correlation coefficients and geometric averages obtained therefrom are rounded to −0.5 for the same reason (reduction of the calculation load and facility of graph drawing on the negative side) as FIGS. 7 through 10. Also in the table of FIG. 14, only 11×11 cells containing a characteristic part of the two-dimensional distribution are shown out of the two-dimensional distribution table having 41×41 cells, for the same reason (limitations of space) as FIG. 9.
  • As shown in FIGS. 14 and 15, the maximum geometric average is 0.003 in the two-dimensional distribution of geometric averages in the second example. In other words, the two-dimensional distribution shows that no part of the subject image has any correlation with the reference image, Meanwhile, it is visually and intuitively obvious that there is almost no correlation between each part of the subject image and the reference image, since the subject image does not contain any color similar to that in the reference image and also does not contain anything having a similar shape to the calyx in the reference image (as is clear from the comparison between FIG. 11 and FIG. 6). Therefore, the two-dimensional distribution of geometric averages is effectively serving as data properly representing the correlation between the subject image and the reference image.
  • FIG. 16 shows a subject image in a third example of the embodiment. In the third example, the subject matter of the subject image is the whole of another pale purple hydrangea at the tip of another spray (different from that of the first example), and the reference image is the same as that of the first example (see FIG. 6). The number of pixels of the subject image shown in FIG. 16 is 64×64 similarly to the first example.
  • FIG. 17 is a three-dimensional graph of the product-moment correlation coefficients regarding the Cr components in the third example, and FIG. 18 is a three-dimensional graph of the product-moment correlation coefficients regarding the Cb components in the third example. FIG. 19 is a table showing the geometric average of the product-moment correlation coefficients at each point in the two-dimensional coordinates, and FIG. 20 is a three-dimensional graph of the geometric averages. Incidentally, also in FIGS. 17 through 20, negative correlation coefficients and geometric averages obtained therefrom are rounded to −0.5 for the aforementioned reason. Also in the table of FIG. 19, only 11×11 cells containing a characteristic part of the two-dimensional distribution are shown out of the two-dimensional distribution table having 41×41 cells for the aforementioned reason.
  • As shown in FIGS. 19 and 20, the maximum geometric average is 0.302 in the two-dimensional distribution of geometric averages in the third example. In other words, the two-dimensional distribution shows that the part of the subject image at the position with the maximum geometric average has a certain degree of correlation with the reference image. This is caused by the existence of a lot of colors in the subject image that are almost the same as those of the reference image.
  • FIG. 21 shows a subject image in a fourth example of the embodiment. In the fourth example, the subject matter of the subject image is the whole of still another pale purple hydrangea at the tip of still another spray (different from those of the first and third examples), and the reference image is the same as that of the first example (see FIG. 6). The number of pixels of the subject image shown in FIG. 21 is 64×64 similarly to the first example.
  • FIG. 22 is a three-dimensional graph of the product-moment correlation coefficients regarding the Cr components in the fourth example, and FIG. 23 is a three-dimensional graph of the product-moment correlation coefficients regarding the Cb components in the fourth example. FIG. 24 is a table showing the geometric average of the product-moment correlation coefficients at each point in the two-dimensional coordinates, and FIG. 25 is a three-dimensional graph of the geometric averages. Also in FIGS. 22 through 25, negative correlation coefficients and geometric averages obtained therefrom are rounded to −0.5. Also in the table of FIG. 24, only 11×11 cells containing a characteristic part of the two-dimensional distribution are shown out of the two-dimensional distribution table having 41×41 cells.
  • As shown in FIGS. 24 and 25, the maximum geometric average is 0.349 in the two-dimensional distribution of geometric averages in the fourth example. In other words, similarly to the third example, the two-dimensional distribution shows that the part of the subject image at the position with the maximum geometric average has a certain degree of correlation with the reference image. This is also caused by the existence of a lot of colors in the subject image that are almost the same as those of the reference image.
  • FIG. 26 shows a subject image in a fifth example of the embodiment, and FIG. 27 shows a reference image in the fifth example. In the fifth example, the subject matter of the subject image is the upper body of a person in pale pink clothes, and the subject matter of the reference image is a single flower which is pale magenta. The number of pixels of the subject image shown in FIG. 26 is 64×64 and that of the reference image shown in FIG. 27 is 24×24.
  • FIG. 28 is a three-dimensional graph of the product-moment correlation coefficients regarding the Cr components in the fifth example, and FIG. 29 is a three-dimensional graph of the product-moment correlation coefficients regarding the Cb components in the fifth example. FIG. 30 is a table showing the geometric average of the product-moment correlation coefficients at each point in the two-dimensional coordinates, and FIG. 31 is a three-dimensional graph of the geometric averages. Also in FIGS. 28 through 31, negative correlation coefficients and geometric averages obtained therefrom are rounded to −0.5. Also in the table of FIG. 30, only 11×11 cells containing a characteristic part of the two-dimensional distribution are shown out of the two-dimensional distribution table having 41×41 cells.
  • As shown in FIGS. 30 and 31, the maximum geometric average is 0.215 in the two-dimensional distribution of geometric averages in the fifth example. In other words, the two-dimensional distribution shows that the part of the subject image at the position with the maximum geometric average has little correlation with the reference image. This result indicates that the correlation does not get high enough only by the existence of colors in the subject image that are similar to those of the reference image.
  • FIG. 32 shows a subject image in a sixth example of the embodiment, and FIG. 33 shows a reference image in the sixth example. In the sixth example, the subject matter of the subject image is the upper body of a person, and the reference image is an image obtained by directly cutting out a part containing the face of the person from the subject image. The number of pixels of the subject image shown in FIG. 32 is 68×70 and that of the reference image shown in FIG. 33 is 34×30.
  • FIG. 34 is a three-dimensional graph of the product-moment correlation coefficients regarding the Cr components in the sixth example, and FIG. 35 is a three-dimensional graph of the product-moment correlation coefficients regarding the Cb components in the sixth example. FIG. 36 is a table showing the geometric average of the product-moment correlation coefficients at each point in the two-dimensional coordinates, and FIG. 37 is a three-dimensional graph of the geometric averages. Also in FIGS. 34 through 37, negative correlation coefficients and geometric averages obtained therefrom are rounded to −0.5. While the table showing the two-dimensional distribution of the geometric averages should have 35×41 cells statistically, only 17×9 cells containing a characteristic part of the two-dimensional distribution are shown in FIG. 36 due to limitations of space.
  • As shown in FIGS. 36 and 37, the maximum geometric average is 1.000 in the two-dimensional distribution of geometric averages in the sixth example. In other words, the two-dimensional distribution shows that the part of the subject image at the position with the maximum geometric average has an extremely high correlation with the reference image. Since the part of the subject image has exactly the same contents as the reference image, it can be said that the two-dimensional distribution of geometric averages is effectively serving as data properly representing the correlation between the subject image and the reference image.
  • FIG. 38 shows a subject image in a seventh example of the embodiment. In the seventh example, the subject matter of the subject image is modern buildings, and the reference image is the same as that of the sixth example (see FIG. 33). The number of pixels of the subject image shown in FIG. 38 is 88×84.
  • FIG. 39 is a three-dimensional graph of the product-moment correlation coefficients regarding the Cr components in the seventh example, and FIG. 40 is a three-dimensional graph of the product-moment correlation coefficients regarding the Cb components in the seventh example. FIG. 41 is a table showing the geometric average of the product-moment correlation coefficients at each point in the two-dimensional coordinates, and FIG. 42 is a three-dimensional graph of the geometric averages. Also in FIGS. 39 through 42, negative correlation coefficients and geometric averages obtained therefrom are rounded to −0.5. Also in the table of FIG. 41, only 24×8 cells containing a characteristic part of the two-dimensional distribution are shown out of the two-dimensional distribution table having 55×55 cells.
  • As shown in FIGS. 41 and 42, the maximum geometric average is 0.232 in the two-dimensional distribution of geometric averages in the seventh example. In other words, similarly to the fifth example, the two-dimensional distribution shows that the part of the subject image at the position with the maximum geometric average has little correlation with the reference image. This is caused by the fact that the subject image has almost no color that is the same as or similar to that in the reference image.
  • FIG. 43 shows a subject image in an eighth example of the embodiment. In the eighth example, the subject matter of the subject image is the upper body of another person (different from that of the sixth example), and the reference image is the same as that of the sixth example (see FIG. 33). The number of pixels of the subject image shown in FIG. 43 is 72×48.
  • FIG. 44 is a three-dimensional graph of the product-moment correlation coefficients regarding the Cr components in the eighth example, and FIG. 45 is a three-dimensional graph of the product-moment correlation coefficients regarding the Cb components in the eighth example. FIG. 46 is a table showing the geometric average of the product-moment correlation coefficients at each point in the two-dimensional coordinates, and FIG. 47 is a three-dimensional graph of the geometric averages. Also in FIGS. 44 through 47, negative correlation coefficients and geometric averages obtained therefrom are rounded to −0.5. Also in the table of FIG. 46, only 19×7 cells containing a characteristic part of the two-dimensional distribution are shown out of the two-dimensional distribution table having 39×19 cells.
  • As shown in FIGS. 46 and 47, the maximum geometric average is 0.315 in the two-dimensional distribution of geometric averages in the eighth example. In other words, the two-dimensional distribution shows that the part of the subject image at the position with the maximum geometric average has a certain degree of correlation with the reference image. This is caused by the similarity of color and shape between the two faces.
  • FIG. 48 shows a subject image in a ninth example of the embodiment, and FIG. 49 shows a reference image in the ninth example. In the ninth example, the subject matter of the subject image is the upper body of still another person (different from those of the sixth and eighth examples), and the reference image is an image obtained by directly cutting out a facial part of another image containing a person that is different from the aforementioned persons. The number of pixels of the subject image shown in FIG. 48 is 64×64 and that of the reference image shown in FIG. 49 is 24×24.
  • FIG. 50 is a three-dimensional graph of the product-moment correlation coefficients regarding the Cr components in the ninth example, and FIG. 51 is a three-dimensional graph of the product-moment correlation coefficients regarding the Cb components in the ninth example. FIG. 52 is a table showing the geometric average of the product-moment correlation coefficients at each point in the two-dimensional coordinates, and FIG. 53 is a three-dimensional graph of the geometric averages. Also in FIGS. 50 through 53, negative correlation coefficients and geometric averages obtained therefrom are rounded to −0.5. Also in the table of FIG. 52, only 20×7 cells containing a characteristic part of the two-dimensional distribution are shown out of the two-dimensional distribution table having 41×41 cells.
  • As shown in FIGS. 52 and 53, the maximum geometric average is 0.355 in the two-dimensional distribution of geometric averages in the ninth example. In other words, the two-dimensional distribution shows that the part of the subject image at the position with the maximum geometric average has a certain degree of correlation with the reference image. This is caused by the similarity of color and shape between the two faces, similarly to the eighth example.
  • As described above, even if the reference image data contains exactly a portion of the subject image data, the type of the subject can be correctly identified. Since the size of the images (reference image and the subject image) are relatively small, the comparison (evaluation) process can be executed within a relatively short period of time. According to experiments, when the subject image data has 64×64 pixels, and the reference image data has 20×20 pixels, the type of the subject can be identified precisely.
  • In the above embodiment, each model image is converted to a reference image having less size than the model image. However, depending on the performance of the CPU 70 a, original size of the model image, the size of the subject image etc., the model image may be used as it is. In such a case, the “reference image data” in the foregoing description should be interpreted equal to the “model image data”.
  • In the above-described embodiment, the reference image data has a lower number of pixels than the subject image data. This may be reversed and the type of the subject can also be identified correctly.
  • In the embodiment, all the model images are compared with the subject image to determine the type of the subject image. This may be modified so that only one or a predetermined number of model images are used to detect the type of the subject image.
  • Incidentally, while the image pickup device 10 has been assumed in the above embodiment to generate the image data constantly, the operation of the image pickup device 10 is not restricted to that in the above explanation. For example, the image pickup device 10 may also be configured to generate the image data only when the shutter button on the control panel 60 is pressed halfway or fully. In this case, the operator (user) can let the display 50 indicate the subject information corresponding to a subject that he/she hopes to shoot, only by aiming the digital still camera at the subject and pressing the shutter button halfway, by which the operator can easily learn which type of image quality correction is going to be executed to the subject to be shot.
  • The two-dimensional distribution data of product-moment correlation coefficients (or the two-dimensional distribution data of geometric averages) used in the above embodiment is information for judging the similarity between each part of the subject image data and each reference image data (or each model image data). By reducing the size (the number of pixels) of the reference image data (or model image data) as much as possible within the extent allowing the judgment of the similarity, the calculation load of the matching between the subject image data and the reference image data (or model image data) can be reduced. By the reduction of the calculation load, the subject information can be updated and the settings for the image quality correction can be changed according to the subject information (the type of the subject) constantly, regardless of whether it is during shooting or not. Therefore, a proper image quality correction process depending on the type of the subject can be executed to the image data (obtained by shooting the subject) quickly.
  • In the above-described embodiment, for evaluating the similarity of the reference image data and the subject image data, the product-moment correlation coefficient γ is introduced. It should be appreciated that the evaluation method need not be limited to one employing the product-moment correlation coefficient, and various methods can be used alternatively or optionally. For example, the evaluation referring to a city-block distance, Euclid distance, weighted Euclid distance, Mahalanobis distance or a similar parameter can be used.
  • FIG. 54 is a flowchart illustrating an image capturing process which is executed when a shutter button of the camera is depressed, and FIG. 55 is a flowchart illustrating an image compensating procedure called in the image capturing process shown in FIG. 54.
  • The image capturing process is started when the shutter button (not shown) of the digital still camera is depressed. In S210, the CPU 70 a retrieves the subject information stored in the prescribed area of the RAM 70 b, and based on the subject information, the CPU 70 a sets an aperture value and a shutter speed in accordance with the subject information.
  • In the embodiment, the correspondence between the subject information (i.e., a type of the subject) and the parameters (e.g., aperture value and shutter speed) are preliminarily determined. The correspondence may be modified/input by a user of the digital still camera.
  • In S220, in accordance with the aperture value and shutter speed set in S210, an image of the subject is captured and temporarily stored in a work area of the image processor 30.
  • In S230, the CPU 70 a calls a subroutine, the image compensating process, which will be described referring to FIG. 55.
  • In this process, in accordance with the type of the subject stored and retrieved from the prescribed area of the RAM 70 b, the captured image of the subject is processed.
  • For example, if the type of the subject is a “landscape”, in S232, the saturation with respect blue components is increased. If the type of the subject is a “portrait”, in S233, the color reproduction of flesh color is enhanced, and the edge enhancement is suppressed. If the type of the subject is a building, in S234, the edge enhancement is suppressed. If the type of the subject does not fall in one of the above, a default processing is applied to the image data.
  • It should be appreciated that the above processing of the image is only an example, and various modifications could be made. Further, although the image processing for the three types of subjects is described, the number/kind of types is not limited to the embodiment. Furthermore, the processing to be applied to the image data may be preliminary determined or set by the user of the digital camera.
  • After the image compensating process is completed, control returns to S240 of FIG. 54, at which the processed image is stored in the storage 50.
  • While the present invention has been described with reference to the particular illustrative embodiment, it is not to be restricted by the above exemplary embodiment but only by the appended claims. For example, while the description of the above embodiment has been given taking a digital still camera as an example, the present invention is also applicable to digital video cameras to achieve similar effects. It is to be appreciated that those skilled in the art can change or modify the embodiment without departing from the scope of the present invention.
  • The present disclosure relates to the subject matter contained in Japanese Patent Application No. 2003-366319, filed on Oct. 27, 2003, which is expressly incorporated herein by reference in its entirety.

Claims (10)

1. A computer program product comprising computer-readable instructions that cause a computer to function as a subject judgment device which comprises:
a storage module which stores at least one piece of reference image data in a storage device associated with subject information which defines a subject in an image represented by the reference image data,
an input module to which subject image data is input,
a distribution data generation module which generates two-dimensional distribution data of product-moment correlation coefficients for at least one piece of reference image data stored in the storage device by calculating the product-moment correlation coefficients between the at least one reference image data and parts of the subject image data,
an identification module which identifies a piece of two-dimensional distribution data having a value greater than a predetermined value,
an output module which reads out the subject information associated with the reference image data corresponding to the identified two-dimensional distribution data from the storage device and which outputs the subject information.
2. A computer program product comprising computer-readable instructions that cause a computer to function as:
a storage module which stores a plurality of pieces of reference image data in a storage device associated with subject information defining subjects in images represented by the respective image data of the plurality of reference image data,
an input module to which subject image data is input,
a distribution data generation module which generates two-dimensional distribution data of product-moment correlation coefficients for each of the plurality of pieces of reference image data stored in the storage device by calculating the product-moment correlation coefficients between each of the plurality of pieces of reference image data and parts of the subject image data,
an identification module which identifies a piece of two-dimensional distribution data having the highest maximum value among all the two-dimensional distribution data generated for all the pieces of the reference image data, and
an output module which reads out the subject information associated with the reference image data corresponding to the identified two-dimensional distribution data from the storage device and which outputs the subject information.
3. The computer program product according to claim 2, wherein:
the distribution data generation module includes:
a color space conversion module which converts color space of the subject image data into YCrCb while converting color space of all the reference image data stored in the storage device into YCrCb,
a coefficient calculation module which calculates product-moment correlation coefficients between Cr components of each reference image data and parts of Cr components of the subject image data and thereby generates two-dimensional distribution data of the product-moment correlation coefficients regarding the Cr components for each reference image data and which calculates product-moment correlation coefficients between Cb components of each reference image data and parts of Cb components of the subject image data and thereby generates two-dimensional distribution data of the product-moment correlation coefficients regarding the Cb components for each reference image data, and
an average calculation module which executes an average calculation process, for calculating each geometric average of the product-moment correlation coefficients regarding the Cr and Cb components at the same two-dimensional coordinates, for each reference image data and thereby generates two-dimensional distribution data of the geometric averages for each reference image data, and
wherein the identification module identifies a piece of two-dimensional distribution data of geometric averages having the highest maximum value among all the two-dimensional distribution data of geometric averages generated for all the reference image data.
4. A digital camera comprising:
a storage unit which stores one or more pieces of reference image data associating each reference image data with subject information which defines a subject in an image formed by the reference image data,
an image pickup device which picks up a subject image formed by an objective optical system and thereby successively generates subject image data,
an image processing unit which changes image quality of an image displayed on a display device according to the subject image data by executing computation to the subject image data using proper correction values,
a distribution data generation unit which generates two-dimensional distribution data of product-moment correlation coefficients for each of the one or more pieces of reference image data stored in the storage unit by calculating the product-moment correlation coefficients between each reference image data and parts of the subject image data,
an identification unit which identifies a piece of two-dimensional distribution data having the highest maximum value among all the two-dimensional distribution data generated for all the reference image data,
a readout unit which reads out the subject information associated with the reference image data corresponding to the identified two-dimensional distribution data from the storage unit, and
a setting alteration unit which alters the correction values used by the image processing unit into correction values corresponding to the subject information read out by the readout unit.
5. A method of identifying a type of a subject image, comprising:
preparing a plurality of pieces of reference image data respectively representing a plurality of different types of images, the number of pixels of the subject image data being different from the number of pixels of each of the reference image data;
scanning, of a piece of reference image data and subject image data, the data having the lower number of pixels within the other of the piece of reference image data and the subject image data to evaluate similarity therebetween at every predetermined scanning position to obtain two dimensional distribution data, the scanning being repeated for each of the plurality of pieces of reference image data to obtain the two dimensional distribution data for each of the plurality of pieces of the reference image data; and
identifying one of the plurality of pieces of the reference image data most similar to the subject image in accordance with the plurality of pieces of the two dimensional distribution data corresponding to the plurality of pieces of the reference image data.
6. The method according to claim 5, wherein the number of the pixels of each piece of the reference image data is smaller than the number of the pixels of the subject image data.
7. The method according to claim 5, wherein the numbers of the pixels of the plurality of pieces of the reference image data are the same.
8. The method according to claim 5, wherein the preparing includes generating each of the plurality of pieces of the reference image data by reducing the number of pixels of a plurality of pieces of predetermined model image data.
9. The method according to claim 5, wherein the preparing includes reducing the number of pixels of an object image to generate the subject image.
10. The method according to claim 5, wherein the identifying identifies one of the plurality of pieces of the reference image data which corresponds to the two-dimensional distribution data having the highest maximum value of any two-dimensional distribution data as the reference image data corresponding to the subject image.
US10/973,617 2003-10-27 2004-10-27 Digital camera having subject judgment function Expired - Fee Related US7605847B2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2003366319A JP2005128956A (en) 2003-10-27 2003-10-27 Subject determining program and digital camera
JP2003-366319 2003-10-27

Publications (2)

Publication Number Publication Date
US20050089218A1 true US20050089218A1 (en) 2005-04-28
US7605847B2 US7605847B2 (en) 2009-10-20

Family

ID=34510232

Family Applications (1)

Application Number Title Priority Date Filing Date
US10/973,617 Expired - Fee Related US7605847B2 (en) 2003-10-27 2004-10-27 Digital camera having subject judgment function

Country Status (2)

Country Link
US (1) US7605847B2 (en)
JP (1) JP2005128956A (en)

Cited By (35)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060044444A1 (en) * 2004-08-30 2006-03-02 Pentax Corporation Digital camera
US20060204110A1 (en) * 2003-06-26 2006-09-14 Eran Steinberg Detecting orientation of digital images using face detection information
US20060204055A1 (en) * 2003-06-26 2006-09-14 Eran Steinberg Digital image processing using face detection information
US20070238954A1 (en) * 2005-11-11 2007-10-11 White Christopher A Overlay image contrast enhancement
WO2008054422A2 (en) * 2006-10-30 2008-05-08 Fotonation Vision Limited Digital image processing using face detection and skin tone information
US20080205712A1 (en) * 2007-02-28 2008-08-28 Fotonation Vision Limited Separating Directional Lighting Variability in Statistical Face Modelling Based on Texture Space Decomposition
US20080292193A1 (en) * 2007-05-24 2008-11-27 Fotonation Vision Limited Image Processing Method and Apparatus
US20090102949A1 (en) * 2003-06-26 2009-04-23 Fotonation Vision Limited Perfecting the Effect of Flash within an Image Acquisition Devices using Face Detection
US7684630B2 (en) 2003-06-26 2010-03-23 Fotonation Vision Limited Digital image adjustable compression and resolution using face detection information
US20100272363A1 (en) * 2007-03-05 2010-10-28 Fotonation Vision Limited Face searching and detection in a digital image acquisition device
US7855737B2 (en) 2008-03-26 2010-12-21 Fotonation Ireland Limited Method of making a digital camera image of a scene including the camera user
US7864990B2 (en) 2006-08-11 2011-01-04 Tessera Technologies Ireland Limited Real-time face tracking in a digital image acquisition device
US7912245B2 (en) 2003-06-26 2011-03-22 Tessera Technologies Ireland Limited Method of improving orientation and color balance of digital images using face detection information
US7916897B2 (en) 2006-08-11 2011-03-29 Tessera Technologies Ireland Limited Face tracking for controlling imaging parameters
US7953251B1 (en) 2004-10-28 2011-05-31 Tessera Technologies Ireland Limited Method and apparatus for detection and correction of flash-induced eye defects within digital images using preview or other reference images
US7962629B2 (en) 2005-06-17 2011-06-14 Tessera Technologies Ireland Limited Method for establishing a paired connection between media devices
US7965875B2 (en) 2006-06-12 2011-06-21 Tessera Technologies Ireland Limited Advances in extending the AAM techniques from grayscale to color images
US8050465B2 (en) 2006-08-11 2011-11-01 DigitalOptics Corporation Europe Limited Real-time face tracking in a digital image acquisition device
US8055067B2 (en) 2007-01-18 2011-11-08 DigitalOptics Corporation Europe Limited Color segmentation
US8155397B2 (en) 2007-09-26 2012-04-10 DigitalOptics Corporation Europe Limited Face tracking in a camera processor
US8213737B2 (en) 2007-06-21 2012-07-03 DigitalOptics Corporation Europe Limited Digital image enhancement with reference images
US8330831B2 (en) 2003-08-05 2012-12-11 DigitalOptics Corporation Europe Limited Method of gathering visual meta data using a reference image
US8345114B2 (en) 2008-07-30 2013-01-01 DigitalOptics Corporation Europe Limited Automatic face and skin beautification using face detection
US8379917B2 (en) 2009-10-02 2013-02-19 DigitalOptics Corporation Europe Limited Face recognition performance using additional image features
US8494286B2 (en) 2008-02-05 2013-07-23 DigitalOptics Corporation Europe Limited Face detection in mid-shot digital images
US8498452B2 (en) 2003-06-26 2013-07-30 DigitalOptics Corporation Europe Limited Digital image processing using face detection information
US8503800B2 (en) 2007-03-05 2013-08-06 DigitalOptics Corporation Europe Limited Illumination detection using classifier chains
US8509496B2 (en) 2006-08-11 2013-08-13 DigitalOptics Corporation Europe Limited Real-time face tracking with reference images
US8593542B2 (en) 2005-12-27 2013-11-26 DigitalOptics Corporation Europe Limited Foreground/background separation using reference images
US8675991B2 (en) 2003-06-26 2014-03-18 DigitalOptics Corporation Europe Limited Modification of post-viewing parameters for digital images using region or feature information
US8682097B2 (en) 2006-02-14 2014-03-25 DigitalOptics Corporation Europe Limited Digital image enhancement with reference images
US8989453B2 (en) 2003-06-26 2015-03-24 Fotonation Limited Digital image processing using face detection information
US9129381B2 (en) 2003-06-26 2015-09-08 Fotonation Limited Modification of post-viewing parameters for digital images using image region or feature information
US9467596B2 (en) * 2014-10-31 2016-10-11 Pfu Limited Image-processing apparatus, image-processing method, and computer program product
US9692964B2 (en) 2003-06-26 2017-06-27 Fotonation Limited Modification of post-viewing parameters for digital images using image region or feature information

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4451929A (en) * 1978-11-10 1984-05-29 Hajime Industries Ltd. Pattern discrimination method
US5469512A (en) * 1992-09-08 1995-11-21 Sony Corporation Pattern recognition device
US20030107653A1 (en) * 2001-12-10 2003-06-12 Advanced Telecommunications Research Institute International Method and apparatus for object recognition
US6876755B1 (en) * 1998-12-02 2005-04-05 The University Of Manchester Face sub-space determination
US6940545B1 (en) * 2000-02-28 2005-09-06 Eastman Kodak Company Face detecting camera and method
US6961465B2 (en) * 2000-12-28 2005-11-01 Canon Kabushiki Kaisha System and method for efficient determination of recognition initial conditions
US7184078B2 (en) * 2001-07-02 2007-02-27 Minolta Co., Ltd. Image-processing apparatus and image-quality control method
US7305122B2 (en) * 2002-08-13 2007-12-04 Lockheed Martin Corporation Method and computer program product for identifying and correcting systematic noise in a pattern recognition system

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3421456B2 (en) * 1994-12-13 2003-06-30 富士通株式会社 Image processing device
JP2000278524A (en) * 1999-03-25 2000-10-06 Konica Corp Image processing method for photographic recording element and image forming processor
US6963666B2 (en) 2000-09-12 2005-11-08 Pentax Corporation Matching device
US6959113B2 (en) 2000-09-29 2005-10-25 Pentax Corporation Arbitrary-shape image-processing device and arbitrary-shape image-reproducing device

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4451929A (en) * 1978-11-10 1984-05-29 Hajime Industries Ltd. Pattern discrimination method
US5469512A (en) * 1992-09-08 1995-11-21 Sony Corporation Pattern recognition device
US6876755B1 (en) * 1998-12-02 2005-04-05 The University Of Manchester Face sub-space determination
US6940545B1 (en) * 2000-02-28 2005-09-06 Eastman Kodak Company Face detecting camera and method
US6961465B2 (en) * 2000-12-28 2005-11-01 Canon Kabushiki Kaisha System and method for efficient determination of recognition initial conditions
US7184078B2 (en) * 2001-07-02 2007-02-27 Minolta Co., Ltd. Image-processing apparatus and image-quality control method
US20030107653A1 (en) * 2001-12-10 2003-06-12 Advanced Telecommunications Research Institute International Method and apparatus for object recognition
US7305122B2 (en) * 2002-08-13 2007-12-04 Lockheed Martin Corporation Method and computer program product for identifying and correcting systematic noise in a pattern recognition system

Cited By (83)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8131016B2 (en) 2003-06-26 2012-03-06 DigitalOptics Corporation Europe Limited Digital image processing using face detection information
US8498446B2 (en) 2003-06-26 2013-07-30 DigitalOptics Corporation Europe Limited Method of improving orientation and color balance of digital images using face detection information
US20060204055A1 (en) * 2003-06-26 2006-09-14 Eran Steinberg Digital image processing using face detection information
US9692964B2 (en) 2003-06-26 2017-06-27 Fotonation Limited Modification of post-viewing parameters for digital images using image region or feature information
US9129381B2 (en) 2003-06-26 2015-09-08 Fotonation Limited Modification of post-viewing parameters for digital images using image region or feature information
US9053545B2 (en) 2003-06-26 2015-06-09 Fotonation Limited Modification of viewing parameters for digital images using face detection information
US8989453B2 (en) 2003-06-26 2015-03-24 Fotonation Limited Digital image processing using face detection information
US8948468B2 (en) 2003-06-26 2015-02-03 Fotonation Limited Modification of viewing parameters for digital images using face detection information
US8675991B2 (en) 2003-06-26 2014-03-18 DigitalOptics Corporation Europe Limited Modification of post-viewing parameters for digital images using region or feature information
US20090102949A1 (en) * 2003-06-26 2009-04-23 Fotonation Vision Limited Perfecting the Effect of Flash within an Image Acquisition Devices using Face Detection
US20060204110A1 (en) * 2003-06-26 2006-09-14 Eran Steinberg Detecting orientation of digital images using face detection information
US7684630B2 (en) 2003-06-26 2010-03-23 Fotonation Vision Limited Digital image adjustable compression and resolution using face detection information
US7693311B2 (en) 2003-06-26 2010-04-06 Fotonation Vision Limited Perfecting the effect of flash within an image acquisition devices using face detection
US7702136B2 (en) 2003-06-26 2010-04-20 Fotonation Vision Limited Perfecting the effect of flash within an image acquisition devices using face detection
US20100165150A1 (en) * 2003-06-26 2010-07-01 Fotonation Vision Limited Detecting orientation of digital images using face detection information
US7809162B2 (en) 2003-06-26 2010-10-05 Fotonation Vision Limited Digital image processing using face detection information
US8498452B2 (en) 2003-06-26 2013-07-30 DigitalOptics Corporation Europe Limited Digital image processing using face detection information
US7844135B2 (en) 2003-06-26 2010-11-30 Tessera Technologies Ireland Limited Detecting orientation of digital images using face detection information
US7844076B2 (en) 2003-06-26 2010-11-30 Fotonation Vision Limited Digital image processing using face detection and skin tone information
US7848549B2 (en) 2003-06-26 2010-12-07 Fotonation Vision Limited Digital image processing using face detection information
US7853043B2 (en) 2003-06-26 2010-12-14 Tessera Technologies Ireland Limited Digital image processing using face detection information
US8326066B2 (en) 2003-06-26 2012-12-04 DigitalOptics Corporation Europe Limited Digital image adjustable compression and resolution using face detection information
US7860274B2 (en) 2003-06-26 2010-12-28 Fotonation Vision Limited Digital image processing using face detection information
US8265399B2 (en) 2003-06-26 2012-09-11 DigitalOptics Corporation Europe Limited Detecting orientation of digital images using face detection information
US8224108B2 (en) 2003-06-26 2012-07-17 DigitalOptics Corporation Europe Limited Digital image processing using face detection information
US7912245B2 (en) 2003-06-26 2011-03-22 Tessera Technologies Ireland Limited Method of improving orientation and color balance of digital images using face detection information
US8126208B2 (en) 2003-06-26 2012-02-28 DigitalOptics Corporation Europe Limited Digital image processing using face detection information
US8055090B2 (en) 2003-06-26 2011-11-08 DigitalOptics Corporation Europe Limited Digital image processing using face detection information
US8005265B2 (en) 2003-06-26 2011-08-23 Tessera Technologies Ireland Limited Digital image processing using face detection information
US8330831B2 (en) 2003-08-05 2012-12-11 DigitalOptics Corporation Europe Limited Method of gathering visual meta data using a reference image
US7508438B2 (en) 2004-08-30 2009-03-24 Hoya Corporation Digital camera having a bracketing capability
US20060044444A1 (en) * 2004-08-30 2006-03-02 Pentax Corporation Digital camera
US20090219430A1 (en) * 2004-08-30 2009-09-03 Hoya Corporation Digital camera having a bracketing capability
US8320641B2 (en) 2004-10-28 2012-11-27 DigitalOptics Corporation Europe Limited Method and apparatus for red-eye detection using preview or other reference images
US7953251B1 (en) 2004-10-28 2011-05-31 Tessera Technologies Ireland Limited Method and apparatus for detection and correction of flash-induced eye defects within digital images using preview or other reference images
US8135184B2 (en) 2004-10-28 2012-03-13 DigitalOptics Corporation Europe Limited Method and apparatus for detection and correction of multiple image defects within digital images using preview or other reference images
US7962629B2 (en) 2005-06-17 2011-06-14 Tessera Technologies Ireland Limited Method for establishing a paired connection between media devices
US20070238954A1 (en) * 2005-11-11 2007-10-11 White Christopher A Overlay image contrast enhancement
US20130083981A1 (en) * 2005-11-11 2013-04-04 Christopher A. White Overlay image contrast enhancement
US8275449B2 (en) * 2005-11-11 2012-09-25 Visualsonics Inc. Overlay image contrast enhancement
US8593542B2 (en) 2005-12-27 2013-11-26 DigitalOptics Corporation Europe Limited Foreground/background separation using reference images
US8682097B2 (en) 2006-02-14 2014-03-25 DigitalOptics Corporation Europe Limited Digital image enhancement with reference images
US7965875B2 (en) 2006-06-12 2011-06-21 Tessera Technologies Ireland Limited Advances in extending the AAM techniques from grayscale to color images
US8050465B2 (en) 2006-08-11 2011-11-01 DigitalOptics Corporation Europe Limited Real-time face tracking in a digital image acquisition device
US20110129121A1 (en) * 2006-08-11 2011-06-02 Tessera Technologies Ireland Limited Real-time face tracking in a digital image acquisition device
US7864990B2 (en) 2006-08-11 2011-01-04 Tessera Technologies Ireland Limited Real-time face tracking in a digital image acquisition device
US8270674B2 (en) 2006-08-11 2012-09-18 DigitalOptics Corporation Europe Limited Real-time face tracking in a digital image acquisition device
US7916897B2 (en) 2006-08-11 2011-03-29 Tessera Technologies Ireland Limited Face tracking for controlling imaging parameters
US8509496B2 (en) 2006-08-11 2013-08-13 DigitalOptics Corporation Europe Limited Real-time face tracking with reference images
US8055029B2 (en) 2006-08-11 2011-11-08 DigitalOptics Corporation Europe Limited Real-time face tracking in a digital image acquisition device
US8385610B2 (en) 2006-08-11 2013-02-26 DigitalOptics Corporation Europe Limited Face tracking for controlling imaging parameters
WO2008054422A3 (en) * 2006-10-30 2008-11-20 Fotonation Vision Ltd Digital image processing using face detection and skin tone information
WO2008054422A2 (en) * 2006-10-30 2008-05-08 Fotonation Vision Limited Digital image processing using face detection and skin tone information
US8055067B2 (en) 2007-01-18 2011-11-08 DigitalOptics Corporation Europe Limited Color segmentation
US20080205712A1 (en) * 2007-02-28 2008-08-28 Fotonation Vision Limited Separating Directional Lighting Variability in Statistical Face Modelling Based on Texture Space Decomposition
US8224039B2 (en) 2007-02-28 2012-07-17 DigitalOptics Corporation Europe Limited Separating a directional lighting variability in statistical face modelling based on texture space decomposition
US8509561B2 (en) 2007-02-28 2013-08-13 DigitalOptics Corporation Europe Limited Separating directional lighting variability in statistical face modelling based on texture space decomposition
US8503800B2 (en) 2007-03-05 2013-08-06 DigitalOptics Corporation Europe Limited Illumination detection using classifier chains
US9224034B2 (en) 2007-03-05 2015-12-29 Fotonation Limited Face searching and detection in a digital image acquisition device
US8923564B2 (en) 2007-03-05 2014-12-30 DigitalOptics Corporation Europe Limited Face searching and detection in a digital image acquisition device
US8649604B2 (en) 2007-03-05 2014-02-11 DigitalOptics Corporation Europe Limited Face searching and detection in a digital image acquisition device
US20100272363A1 (en) * 2007-03-05 2010-10-28 Fotonation Vision Limited Face searching and detection in a digital image acquisition device
US7916971B2 (en) 2007-05-24 2011-03-29 Tessera Technologies Ireland Limited Image processing method and apparatus
US8515138B2 (en) 2007-05-24 2013-08-20 DigitalOptics Corporation Europe Limited Image processing method and apparatus
US20110235912A1 (en) * 2007-05-24 2011-09-29 Tessera Technologies Ireland Limited Image Processing Method and Apparatus
US20110234847A1 (en) * 2007-05-24 2011-09-29 Tessera Technologies Ireland Limited Image Processing Method and Apparatus
US20080292193A1 (en) * 2007-05-24 2008-11-27 Fotonation Vision Limited Image Processing Method and Apparatus
US8494232B2 (en) 2007-05-24 2013-07-23 DigitalOptics Corporation Europe Limited Image processing method and apparatus
US8213737B2 (en) 2007-06-21 2012-07-03 DigitalOptics Corporation Europe Limited Digital image enhancement with reference images
US8896725B2 (en) 2007-06-21 2014-11-25 Fotonation Limited Image capture device with contemporaneous reference image capture mechanism
US10733472B2 (en) 2007-06-21 2020-08-04 Fotonation Limited Image capture device with contemporaneous image correction mechanism
US9767539B2 (en) 2007-06-21 2017-09-19 Fotonation Limited Image capture device with contemporaneous image correction mechanism
US8155397B2 (en) 2007-09-26 2012-04-10 DigitalOptics Corporation Europe Limited Face tracking in a camera processor
US8494286B2 (en) 2008-02-05 2013-07-23 DigitalOptics Corporation Europe Limited Face detection in mid-shot digital images
US20110053654A1 (en) * 2008-03-26 2011-03-03 Tessera Technologies Ireland Limited Method of Making a Digital Camera Image of a Scene Including the Camera User
US8243182B2 (en) 2008-03-26 2012-08-14 DigitalOptics Corporation Europe Limited Method of making a digital camera image of a scene including the camera user
US7855737B2 (en) 2008-03-26 2010-12-21 Fotonation Ireland Limited Method of making a digital camera image of a scene including the camera user
US8384793B2 (en) 2008-07-30 2013-02-26 DigitalOptics Corporation Europe Limited Automatic face and skin beautification using face detection
US9007480B2 (en) 2008-07-30 2015-04-14 Fotonation Limited Automatic face and skin beautification using face detection
US8345114B2 (en) 2008-07-30 2013-01-01 DigitalOptics Corporation Europe Limited Automatic face and skin beautification using face detection
US8379917B2 (en) 2009-10-02 2013-02-19 DigitalOptics Corporation Europe Limited Face recognition performance using additional image features
US10032068B2 (en) 2009-10-02 2018-07-24 Fotonation Limited Method of making a digital camera image of a first scene with a superimposed second scene
US9467596B2 (en) * 2014-10-31 2016-10-11 Pfu Limited Image-processing apparatus, image-processing method, and computer program product

Also Published As

Publication number Publication date
JP2005128956A (en) 2005-05-19
US7605847B2 (en) 2009-10-20

Similar Documents

Publication Publication Date Title
US7605847B2 (en) Digital camera having subject judgment function
US8111881B2 (en) Image pickup device, face detection method, and computer-readable recording medium
EP1686787B1 (en) Image pickup apparatus with special effect mode and method thereof
US8836819B2 (en) Method and apparatus for automatically categorizing images in a digital camera
US20020061142A1 (en) Image correction apparatus
US20190335077A1 (en) Systems and methods for image capture and processing
JP2006080746A (en) Image processor, electronic camera, and image processing program
CN101155263A (en) Image processing apparatus, image processing method, image processing program, and image pickup apparatus
US20020080251A1 (en) Digital imaging device, image processing device, digital imaging method, and image processing program
CN112019739A (en) Shooting control method and device, electronic equipment and storage medium
US20060139669A1 (en) Image pickup apparatus and control method of the apparatus
CN113014803A (en) Filter adding method and device and electronic equipment
US20060274180A1 (en) Image display apparatus
CN101715065A (en) Image correction apparatus, and image correction program
JP5414216B2 (en) Imaging apparatus, control method thereof, and program
JP4632417B2 (en) Imaging apparatus and control method thereof
JP2010147660A (en) Image processor, electronic camera and image processing program
JP4370146B2 (en) Image processing program, image processing method, and recording medium
JP4885079B2 (en) Digital camera, photographing method and photographing program
US7965889B2 (en) Imaging apparatus, imaging method, program and recording medium
JP2006148326A (en) Imaging apparatus and method of controlling the same
JP2010050602A (en) Photographing device, image processing method, and program
US20120154617A1 (en) Image capturing device
CN111476731A (en) Image correction method, image correction device, storage medium and electronic equipment
JP2003209857A (en) Image processing apparatus

Legal Events

Date Code Title Description
AS Assignment

Owner name: PENTAX CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:CHIBA, TORU;REEL/FRAME:015935/0485

Effective date: 20041027

AS Assignment

Owner name: HOYA CORPORATION, JAPAN

Free format text: MERGER;ASSIGNOR:PENTAX COPORATION;REEL/FRAME:022865/0559

Effective date: 20080331

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

AS Assignment

Owner name: PENTAX RICOH IMAGING COMPANY, LTD., JAPAN

Free format text: CORPORATE SPLIT;ASSIGNOR:HOYA CORPORATION;REEL/FRAME:027315/0115

Effective date: 20111003

REMI Maintenance fee reminder mailed
LAPS Lapse for failure to pay maintenance fees
STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20131020