US20060182433A1 - Electronic camera - Google Patents

Electronic camera Download PDF

Info

Publication number
US20060182433A1
US20060182433A1 US11/345,393 US34539306A US2006182433A1 US 20060182433 A1 US20060182433 A1 US 20060182433A1 US 34539306 A US34539306 A US 34539306A US 2006182433 A1 US2006182433 A1 US 2006182433A1
Authority
US
United States
Prior art keywords
face
shooting
area
electronic camera
focus area
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/345,393
Inventor
Takumi Kawahara
Yasuyuki Motoki
Akira Ohmura
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nikon Corp
Original Assignee
Nikon Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from JP2005037675A external-priority patent/JP4639837B2/en
Priority claimed from JP2005038424A external-priority patent/JP4581730B2/en
Application filed by Nikon Corp filed Critical Nikon Corp
Assigned to NIKON CORPORATION reassignment NIKON CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: KAWAHARA, TAKUMI, MOTOKI, YASUYUKI, OHMURA, AKIRA
Publication of US20060182433A1 publication Critical patent/US20060182433A1/en
Priority to US12/289,747 priority Critical patent/US7881601B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03BAPPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
    • G03B3/00Focusing arrangements of general interest for cameras, projectors or printers
    • G03B3/10Power-operated focusing
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03BAPPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
    • G03B17/00Details of cameras or camera bodies; Accessories therefor
    • G03B17/18Signals indicating condition of a camera member or suitability of light
    • G03B17/20Signals indicating condition of a camera member or suitability of light visible in viewfinder
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/61Control of cameras or camera modules based on recognised objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/61Control of cameras or camera modules based on recognised objects
    • H04N23/611Control of cameras or camera modules based on recognised objects where the recognised objects include parts of the human body
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders
    • H04N23/631Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
    • H04N23/632Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters for displaying or modifying preview images prior to image capturing, e.g. variety of image resolutions or capturing parameters
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders
    • H04N23/633Control of cameras or camera modules by using electronic viewfinders for displaying additional information relating to control or operation of the camera
    • H04N23/635Region indicators; Field of view indicators
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/67Focus control based on electronic image sensor signals
    • H04N23/673Focus control based on electronic image sensor signals based on contrast or high frequency components of image signals, e.g. hill climbing method
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N2101/00Still video cameras

Definitions

  • the present invention relates to an electronic camera which captures an image of a subject, and particularly relates to an electronic camera which can detecting a characteristic portion of a subject, for example, a face.
  • Japanese Unexamined Patent Application Publication No. 2001-215403 discloses an electronic camera having a face recognition function and performing focusing control according to the eyes of a subject.
  • the electronic camera in Japanese Unexamined Patent Application Publication No. 2001-215403 has a difficulty in focusing on the eyes of the subject when the subject closes his or her eyes or wears glasses, and the focusing operation thereof is low in stability. Therefore, there is still room for improvement.
  • An object of the present invention is to provide an electronic camera which prevents a user from being confused.
  • Another object of the present invention is to provide an electronic camera which can stably focus on a person as a subject by recognizing his/her face.
  • the scene shooting mode for shooting the scene including a person be a portrait shooting mode.
  • the controlling section does not allow the face detecting section to perform the face detection when the portrait shooting mode is a portrait shooting mode for shooting a night landscape.
  • the controlling section control a shooting lens to focus on a face area detected by the face detecting section.
  • the electronic camera of the above first aspect it is desirable that it further include a function setting section which sets a function for each scene shooting mode, and that the scene shooting mode for shooting the scene including a person is provided with a setting item regarding the face detection.
  • the electronic camera of the above first aspect it is desirable that it stop a digital zoom function of electronically magnifying a magnifying power during the face detection by the face detecting section.
  • the electronic camera of the above first aspect it is desirable that it stop a closeup shooting function of shifting a shooting lens for closeup shooting during the face detection by the face detecting section.
  • the electronic camera of the above first aspect it is desirable that it further include a display section which displays a subject image obtained before shooting, and that during the face detection by the face detecting section, the amount of shooting information for display on the display section is reduced compared with while the face detection is not performed.
  • the focus area specifying section change the specified focus area to a focus area located below the face area, when the focusing position is not detected in the focus area including the contour of the face area.
  • the face recognizing section detects a direction of the face based on a positional relation of face parts in the face area. It is desirable that the focus area specifying section change a position of the focus area to be the specified focus area, according to the direction of the face.
  • the electronic camera of the above second aspect it is desirable that it further include an attitude detecting section which detects a shooting attitude of the electronic camera, and that the focus area specifying section change a position of the focus area to be the specified focus area, according to the shooting attitude.
  • the electronic viewfinder displays a viewfinder image of the image shooting plane according to the image signal, and displays an indication of focusing failure associated with a face area of the viewfinder image when the focusing position is not detected in the specified focus area.
  • FIG. 1 is an external view of an electronic camera 1 according to a first embodiment
  • FIG. 2 is a block diagram showing functions of the electronic camera 1 according to the first embodiment
  • FIG. 3 is views showing a mode select dial 105 for selecting a shooting scene mode and a menu screen corresponding to the selected shooting scene mode;
  • FIG. 4 is views showing the mode select dial 105 for selecting a shooting scene mode and a menu screen corresponding to the selected shooting scene mode;
  • FIG. 5 is a flowchart showing control performed by a CPU 111 ;
  • FIG. 6 is a flowchart showing control performed by the CPU 111 in a face recognizing AF mode
  • FIG. 7 is a flowchart showing face detection AF control performed by the CPU 111 ;
  • FIG. 8 is a flowchart showing AF control by a detection area performed by the CPU 111 ;
  • FIG. 9 is a flowchart showing AF control, in which a central area is weighted, performed by the CPU 111 ;
  • FIG. 10 is a view showing a display example of an image plane displayed on a monitor 103 ;
  • FIG. 11 is a view showing a display example of the image plane displayed on the monitor 103 ;
  • FIG. 12 is a view showing a display example of the image plane displayed on the monitor 103 ;
  • FIG. 13 is a view showing a display example of the image plane displayed on the monitor 103 ;
  • FIG. 14 is a view showing a display example of the image plane displayed on the monitor 103 ;
  • FIG. 15 is a block diagram showing an overview of an electronic camera of a second embodiment
  • FIG. 16 is a flowchart showing a shooting operation in the second embodiment
  • FIG. 17 is a view showing the position of a specified focus area in the second embodiment
  • FIG. 18 is a view showing a viewfinder image at the time of face recognition in the second embodiment
  • FIG. 19 is a block diagram showing an overview of an electronic camera of a third embodiment
  • FIG. 20 is a flowchart showing a shooting operation in a fourth embodiment.
  • FIG. 21 is a view showing the position of a specified focus area in the fourth embodiment.
  • FIG. 1 is an external view of the electronic camera 1 according to the first embodiment.
  • the electronic camera 1 includes a release button 101 , a cruciform key 102 , a monitor 103 , a decision button 104 , a mode select dial 105 , a zoom button, 106 , a menu button 107 , a play button 108 , a closeup shooting button 109 , and an optical viewfinder 110 .
  • the release button 101 is a button capable of detecting two-stage operations: a half-press stage and a full-press stage.
  • the release button 101 is manipulated by a user when the user instructs the start of shooting.
  • the cruciform key 102 is manipulated by the user to move a cursor or the like on the monitor 103 .
  • the decision button 104 is a button manipulated by the user when the user selects and decides an item with the cruciform key 102 or the like.
  • the decision button 104 is manipulated by the user also when the user switches on/off states of the monitor 103 .
  • the mode select dial 105 is a dial which enables the user to change a camera function such as a shooting scene selection by turning it.
  • the zoom button 106 is a button manipulated by the user when the user optically and electronically scales up or down an image recorded at the time of shooting.
  • the zoom button 106 is manipulated by the user also when at the time of replay of image data, the user electronically scales up or down replayed image.
  • the menu button 107 is a button manipulated by the user when the user wants to display a menu screen.
  • the play button 108 is a button manipulated by the user when the user replays image data recorded in a memory.
  • the closeup shooting button 109 is a button manipulated by the user when the user shoots a close subject such as a close plant.
  • the optical viewfinder 110 is for the user to optically check a field.
  • FIG. 2 is a block diagram showing functions of the electronic camera 1 according to the first embodiment.
  • the electronic camera 1 is composed of a CPU 111 , a face detecting section 112 , a built-in memory 113 , a manipulation circuit 114 , a display circuit 115 , a monitor 103 , an image processing circuit 116 , an image pickup device 117 , a zoom lens 118 , a zoom driver 119 , a focus lens 120 , a focus driver 121 , an aperture 122 , an aperture driver 123 , and a memory card slot 124 .
  • there are other circuits to realize functions of the electronic camera but have little relation to the first embodiment, so that a description thereof will be omitted.
  • the CPU 111 is a circuit which processes programs to realize various functions executed in the electronic camera 1 .
  • the CPU 111 executes the programs stored in a memory in the CPU 111 , that is, the built-in memory 113 and controls various circuits in the electronic camera 1 .
  • the face detecting section 112 extracts a characteristic portion of image data picked up by the image pickup device 117 and detects a face area, face size, and so on of a subject.
  • a function block called the face detecting section 112 is described for explanation, but in the first embodiment, a face detection function is realized in software by a face detection program executed by the CPU 111 .
  • the built-in memory 113 is a memory to store image data, a control program, and so on.
  • a nonvolatile semiconductor memory is used as the built-in memory 113 .
  • the built-in memory 113 stores the face detection program which is executed to detect the face area of the subject. Further, the built-in memory 113 can store face information such as the face position and face size obtained by face detection.
  • the manipulation circuit 114 detects manipulations of manipulation buttons such as the release button 101 , the cruciform key 102 , and the decision button 104 provided in the electronic camera 1 and transfers them to the CPU 111 . Further, the manipulation circuit 114 detects a half-press manipulation and a full-press manipulation of the release button 101 .
  • the display circuit 115 is a circuit to generate image plane data displayed on the monitor 103 .
  • the monitor 103 is a liquid crystal display provided on a rear surface of the electronic camera 1 . This monitor 103 displays the image plane data generated by the display circuit 115 .
  • a shooting lens is an optical lens to focus a subject image onto a light-receiving plane of the image pickup device 117 .
  • This shooting lens is composed of the zoom lens 118 , the focus lens 120 , and so on.
  • the zoom lens 118 is a lens to realize scale-up and scale-down of the optical image focused on the image pickup device 117 .
  • This zoom lens 118 is moved by a motor.
  • the zoom driver 119 is a circuit to drive the motor by a command of the CPU 111 and move the zoom lens 119 to a predetermined position.
  • the focus lens 120 is a lens to adjust focus.
  • This focus lens 120 is moved by a motor.
  • the focus driver 121 is a circuit to drive the motor by a command of the CPU 111 and move the focus lens 120 to a predetermined position.
  • the aperture 122 is to adjust the amount of light of the field incident on the image pickup device 117 .
  • the aperture driver 123 is a circuit to drive a motor by a command of the CPU 111 and control open and closed states of the aperture 122 .
  • the image pickup device 117 is a device to convert the optical image inputted through the shooting lens into electric image signals.
  • the image pickup device 118 is composed of a CCD, or the like.
  • the image processing circuit 116 analog-to-digital converts the electric signals outputted from the image pickup device 117 to generate digital signals. Further, the image processing circuit 116 performs interpolation processing or the like on the digital converted signals to generate image data.
  • a memory card is inserted into the memory card slot 124 .
  • the memory card slot 124 writes data such as image data in the memory card or deletes data in the memory card.
  • a setting manipulation to use the face detection function in the electronic camera 1 will be described.
  • the main subject is a person
  • the user of the camera wants to focus on a face portion obtained by face detection.
  • a result of face detection is used as one of options for deciding an area brought into autofocus (AF).
  • FIG. 3 and FIG. 4 are views each showing the mode select dial 105 to select a shooting scene mode and a menu screen corresponding to the selected shooting scene mode.
  • FIG. 3A shows the mode select dial 105 in a state where a portrait shooting mode is selected.
  • FIG. 3B shows a menu screen displayed on the monitor 103 when the menu button 107 is manipulated in the portrait shooting mode. The user selects a function icon by manipulating the cruciform key 102 on the menu screen and manipulates the decision button 104 . Thus, the electronic camera 1 performs a selected function.
  • An icon 201 in FIG. 3B is an icon selected when a face detection AF function is performed. The operation of face detection AF will be described later.
  • FIG. 4A shows the mode select dial 105 in a state where a night landscape portrait shooting mode is selected.
  • FIG. 4B shows a menu screen displayed on the monitor 103 when the menu button 107 is manipulated in the night landscape portrait shooting mode.
  • the menu screen of the night landscape portrait mode is not provided with the icon 201 which is selectable from the menu screen of the portrait mode.
  • a person is sometimes shot in the night landscape portrait mode. Therefore, it is effective to detect a face even in the night landscape portrait mode and autofocus on a detected face area.
  • some degree of brightness of a face portion is required, but it is expected that the brightness is not sufficient to detect the face in a scene of the night landscape portrait.
  • the face detection AF is made unsettable in the night landscape portrait mode.
  • the sport shooting mode is not provided with the icon to select the face detection AF either. This results in avoiding in advance the user from being confused because the user cannot recognize the face.
  • the electronic camera 1 When the face detection AF of the icon 201 on the menu screen shown in FIG. 3B is selected, the electronic camera 1 performs control so as to focus on the closest face portion of the detected face. When the face detection AF is selected, the electronic camera 1 is automatically switched to a constant AF mode. In the constant AF mode, irrespective of the manipulation of the release button 101 , the electronic camera 1 repeats focusing by AF. Then, by half pressing the release button 101 , the electronic camera 1 performs control to make an AF lock.
  • the face detection is performed using a moving image (a through image) picked up by the image pickup device 117 .
  • This through image is displayed on the monitor 103 for check-up of a subject image to be shot.
  • the electronic camera 1 performs control in such a manner as to increase the brightness of the through image as compared with while the face detection is not performed. Increasing the brightness of the through image makes it easy to detect the face.
  • the electronic camera 1 repeatedly performs the face detection using the through image until the release button 101 is half pressed.
  • the face detection is assisted by performing central-area AF and multi-area AF to bring the subject into focus.
  • the electronic camera 1 repeatedly stores face detection information such as the face position and face size as a result of the face detection in the built-in memory 113 while overwriting the latest piece thereof until the release button 101 is half pressed. This makes it possible to read and use the preceding detection result stored in the built-in memory 113 even if the face is not detected through the half-press manipulation. Besides, when the face detection information cannot be obtained in the half-press manipulation, the electronic camera 1 forcibly performs AF using another area such as a central area.
  • FIG. 5 is a flowchart showing control performed by the CPU 111 .
  • the flow shown in FIG. 5 is set to a still image shooting mode, and starts by detecting the manipulation of the menu button 107 .
  • step S 01 the CPU 111 detects whether a scene shooting mode in which it is assumed that a person is shot is selected. In the electronic camera 1 of the first embodiment, it is detected whether the portrait mode or the night landscape portrait mode is selected by the mode select dial 105 . If it is selected, the CPU 111 goes to step S 02 . On the other hand, if it is not selected, the CPU 111 goes to step S 07 . In step S 02 , the CPU 111 determines whether the selected shooting scene is the night landscape portrait mode. If it is the nigh landscape portrait mode, the CPU goes to step S 07 . On the other hand, if it is not the night landscape portrait mode, the CPU goes to step S 03 .
  • step S 03 the menu screen having the face detection AF as an option which is shown in FIG. 3B is displayed on he monitor 103 .
  • step S 04 the CPU 111 detects whether the decision button 104 is manipulated. If the manipulation of the decision button 104 is detected, the CPU 111 goes to step SOS. On the other hand, if the manipulation thereof is not detected, the CPU 111 continues the detection.
  • step 505 the CPU 111 determines whether the face detection AF is selected. If the face detection AF is selected, the CPU 111 goes to step S 06 . On the other hand, if any option other than the face detection AF is selected, the CPU 111 goes to step S 09 .
  • step S 06 the CPU 111 controls the electronic camera 1 in a face recognition AF mode. The face recognition AF mode will be described using a flowchart in FIG. 6 .
  • step S 07 the menu screen shown in FIG. 4B is displayed on the monitor 103 .
  • step S 08 the CPU 111 detects whether the decision button 104 is manipulated. If the manipulation thereof is detected, the CPU 111 goes to step S 09 . On the other hand, if the manipulation thereof is not detected, the CPU 111 continues the detection. Then, in step S 09 , the CPU 111 controls the electronic camera 1 in a normal AF mode.
  • the face detection is prevented from performed in the night landscape mode in which it is probable that the face cannot be detected, even during the scene shooting mode for shooting a person, which prevents the user from being confused.
  • FIG. 6 is the flowchart showing control performed by the CPU 111 in the face recognition AF mode. This flow is executed in step S 06 in FIG. 5 .
  • step S 51 the CPU 111 makes a setting not to perform an electronic zoom function.
  • the face detection can be continued using the through image.
  • step S 52 the CPU 111 makes a setting not to perform a closeup shooting function. This is because through the closeup shooting only part of the face is shot and the part of the face is not sufficient to detect the face.
  • step S 53 the CPU 111 prohibits the monitor 103 from turning off. Without the monitor 103 being in an on state, the user cannot check up the result of face detection, so that he/she takes a shot even when the face detection of the electronic camera 1 is erroneous.
  • step S 54 the CPU 111 switches the display of the monitor 103 to a simple display.
  • the monitor 103 displays a reduced number of items of shooting information such as a memory remaining capacity indication which is superimposed on image data.
  • a display example of shooting information in the face detection AF mode is shown in FIG. 11 .
  • a display example of shooting information in the normal AF mode is shown in FIG. 14 .
  • Switching to the simple display such as shown in FIG. 11 can make it easier for the user to check up a face detection frame as much as possible. This is because in the simple display, it is unlikely that the frame displayed when the face is detected overlaps with the display of the shooting information.
  • the CPU 111 stops the shooting information display. When the shooting information display is not set, it is needless to say that the CPU 111 continues the non-display of the shooting information.
  • step S 55 the CPU 111 measures the brightness of the field using the image captured by the image pickup device 117 .
  • step S 56 the CPU 111 adjusts the brightness based on the measured field brightness. Then, the CPU 111 displays the through image on the monitor 103 .
  • step S 57 the CPU 111 performs face detection AF control.
  • the above control can reduce defects in the face detection AF mode and user's dissatisfaction.
  • FIG. 7 is a flowchart showing the face detection AF control performed by the CPU 111 . This flow is executed in step S 57 in FIG. 6 .
  • step S 101 the CPU 111 determines whether the face is detected. If the face is detected, the CPU 111 goes to step S 102 . On the other hand, if the face is not detected, the CPU 111 goes to step S 106 .
  • step S 102 if a face character shown in FIG. 10 is displayed while being superimposed on the subject image on the monitor 103 , the CPU 111 deletes the display of the face character.
  • step S 103 the CPU 111 sets an area depending on the position and size of the detected face, and as shown in FIG. 11 , displays frames while superimposing them on the subject image on the monitor 103 . If the number of detected faces is plural, a frame is displayed on each face.
  • step S 104 the CPU 111 performs AF in the set area.
  • step S 105 the CPU 111 temporarily stores detected face information in the built-in memory 113 .
  • the CPU 111 can decide an AF area by using the face information stored in the built-in memory 113 when the face is not detected in the half-press manipulation in which the AF area is finally decided.
  • the face information stored in the built-in memory 113 is overwritten with new face information in the next face detection.
  • step S 106 the CPU 111 displays the face character shown in FIG. 10 while superimposing it on the subject image on the monitor 103 .
  • the display of the face character indicates the user that the face detection AF mode is being performed, and informs the user of the size of the face ideal for the face detection control by the size of the face character. Incidentally, it is not necessary to display the face character constantly, and it is only required to display it every several seconds.
  • step S 107 the CPU 111 performs AF in an AF area in which a central area is weighted.
  • step S 108 the CPU 111 determines again whether the face is detected. If the face is not detected, the CPU 111 goes to step S 109 . On the other hand, if the face is detected, the CPU goes to step S 102 . In step S 109 , the CPU 111 performs multi-area AF. Similarly to AF in the central area in step S 107 , a face of a person outside the central portion can be detected.
  • step S 110 the CPU 111 determines again whether the face is detected. If the face is not detected, the CPU 111 goes to step S 111 . On the other hand, if the face is detected, the CPU 111 goes to step S 102 . In step S 111 , the CPU 111 detects whether the release button 101 is half pressed. If the half-press manipulation is detected, the CPU 111 goes to step S 113 . On the other hand, if the half-press manipulation is not detected, the CPU 111 goes to step S 112 . In step S 112 , the CPU 111 deletes the face information stored in the built-in memory 113 and returns to step S 101 .
  • step S 113 if the face character is displayed while being superimposed on the subject image on the monitor 103 , the CPU 111 deletes the display of the face character. Then, in step S 114 , the CPU 111 determines whether the face is detected in order to specify the face which is regarded as a final AF area. If the face is detected, the CPU 111 goes to step S 115 . On the other hand, if the face is not detected, the CPU 111 goes to step S 116 . In step S 115 , the CPU 111 sets an area set according to the position and size of the detected face as the final AF area, and performs AF control. This AF control in the area detected by the face detection will be described later using FIG. 8 .
  • step S 116 the CPU 111 detects whether there is the face information stored in step S 107 in the memory. With the face information therein, the CPU 111 goes to step S 117 . On the other hand, without the face information therein, the CPU 111 goes to step S 118 .
  • step S 117 the CPU 111 sets an area based on the stored face information as the AF area and performs AF control.
  • the AF control in the area detected by the stored face detection will be described later using FIG. 8 . Therefore, if the face is not detected in the half-press manipulation, the face area detected immediately therebefore with little time difference is used as the AF area. This makes it possible to focus on the face portion almost without fail. In particular, this can respond to one-press manipulation in which the release button 101 is fully-pressed with one press.
  • step S 118 the CPU 111 performs AF control with the central area as the AF area. The AF control in the central area will be described later using FIG. 9 .
  • the central area which is likely to include the main subject is automatically used as the AF area. This increases a possibility that the main subject comes into focus. Moreover, it becomes unnecessary for the user to return to the menu screen and reset the AF area, which prevents a photo opportunity from being missed.
  • step S 119 the CPU 111 detects whether the release button 101 is fully pressed. If the release button 101 is fully pressed, the CPU 111 goes to step S 120 . On the other hand, if it is not fully pressed, the CPU 111 goes to step S 121 . In step S 120 , the CPU 111 performs shooting and recording processing. In step S 121 , the CPU 111 detects whether the release button 101 is half pressed. If the release button 101 is half pressed, the CPU 111 returns to step S 119 . On the other hand, if it is not half pressed, the CPU 111 returns to step S 101 .
  • FIG. 8 is a flowchart showing the AF control in the detected area performed by the CPU 111 . This flow is executed in step S 15 and step S 117 in FIG. 7 .
  • step S 201 the CPU 111 displays a face detected area frame set corresponding to the position and size of the detected face while superimposing it on the through image as shown in FIG. 12 .
  • frames are displayed on the respective faces.
  • the frame of the largest or closest face is switched from a white frame (a thin-line frame in FIG. 11 ) to a red frame (a thick-line frame in FIG. 12 ).
  • step S 202 the CPU 111 sets the area displayed by the red frame in FIG. 12 to the AF area and performs AF control.
  • step S 203 the CPU 111 detects whether focus is achieved. If focus is achieved, the CPU 111 goes to step S 204 .
  • step S 204 the CPU 111 switches the frame shown by the thick-line frame in FIG. 12 from the red frame to a blue frame. This makes it possible to inform the user that focus is achieved.
  • a frame display which is not to the AF area may be deleted.
  • FIG. 205 the CPU 111 displays the frame shown by the thick-line frame in FIG. 12 which remains the red frame in a blinking state. This makes it possible to inform the user that focus is not achieved.
  • FIG. 9 is a flowchart showing the AF control in which the central area is weighted performed by the CPU 111 . This flow is executed in step S 118 in FIG. 7 .
  • step S 301 the CPU 111 performs the AF control with the central area being weighted. Then, in step S 302 , the CPU 111 detects whether focus is achieved. If focus is achieved, the CPU 111 goes to step S 303 . On the other hand, if focus is not achieved, the CPU 111 goes to step S 304 .
  • step S 303 the CPU 111 displays an icon 301 shown in FIG. 14 in an on-state while superimposing it on the subject image on the monitor 103 . This makes it possible to inform the user that focus is achieved in an area other than the face detected area.
  • step S 304 the CPU 111 displays the icon 301 shown in FIG. 14 in a blinking state while superimposing it on the subject image on the monitor 103 . This makes it possible to inform the user that focus is not achieved in the area other than the face detected area.
  • FIG. 15 is a block diagram showing an overview of an electronic camera of a second embodiment.
  • the electronic camera of the second embodiment includes a shooting lens 11 , lens driving mechanisms 12 , an image pickup device 13 , an analog signal processing section 14 , an A/D conversion section 15 , an image processing section 16 , a compression/decompression section 17 , a memory 18 , a card I/F 19 , a monitor I/F 20 and a liquid crystal display 21 , a manipulation section 22 , a CPU 23 , and a bus 24 .
  • the image processing section 16 , the compression/decompression section 17 , the memory 18 , the card I/F 19 , the monitor I/F 20 , and the CPU 23 are connected respectively via the bus 24 .
  • the shooting lens 11 is composed of a group of plural lenses including a focusing lens for adjusting the focusing position.
  • the position of this shooing lens 11 in an optical axis direction is adjusted by the lens driving mechanisms 12 .
  • the image pickup device 13 is placed on the image space side of the shooting lens 11 .
  • Photodetectors which photoelectrically convert the subject image to generate analog image signals are two-dimensionally arranged on a light-receiving plane (a plane facing the shooting lens 11 ) of the image pickup device 13 .
  • An output of the image pickup device 13 is connected to the analog signal processing section 14 .
  • the image pickup device 13 exposes the subject at predetermined intervals and outputs the analog image signals (through image signals) by thinning-out reading or the like.
  • This through image signal is used for AF calculation, AE calculation, and face recognition by the CPU 23 , generation of a viewfinder moving image by the image processing section 26 , and so on.
  • the image pickup device 13 of the second embodiment may adopt either a sequential charge transfer method (for example, a CCD) or an XY address method (for example, a CMOS).
  • the analog signal processing section 14 is composed of a CDS circuit which performs correlated double sampling, a gain circuit which amplifies the outputs of the analog image signals, a clamp circuit which clamps the waveform of an input signal at a fixed voltage level, and so on.
  • the A/D conversion section 15 converts the analog image signals outputted from the analog image processing section 14 into digital image signals. An output of the A/D conversion section 15 is connected to the image processing section 16 and the CPU 23 , respectively.
  • the image processing section 16 performs image processing (defective pixel correction, gamma correction, interpolation, color conversion, edge enhancement, and so on) on the digital image signals when the shutter is released to generate shooting image data. Further, the image processing section 16 generates viewfinder images sequentially based on the digital image signals (through image signals) when the shutter is not released.
  • the image processing section 16 combines and displays a rectangular AF frame showing a face area as an AF target with the viewfinder image based on face recognition information described later (See FIG. 18 ). Besides, the image processing section 16 gives an indication of focusing failure to the viewfinder image using the above AF frame based on focusing failure information described later. Examples of this indication of focusing failure are a way of displaying the AF frame in a blinking state, a way of making a change to the color of the AF frame in a normal state, and so on. Incidentally, when the AF calculation is made twice and each ends in focusing failure as described later, the image processing section 16 gives different indications of focusing failure respectively for the first and second times.
  • the compression/decompression section 17 performs processing of compressing the shooting image data after image processing in a JPEG format and processing of decompressing and reconstructing the image data compressed in the JPEG format.
  • the memory 18 is composed of an SDRAM or the like and has a capacity capable of recording image data corresponding to plural frames. Image data before and after the image processing by the image processing section 16 is temporarily stored in this memory 18 .
  • a connector to connect storage media 25 is formed in the card I/F 19 .
  • the storage media 25 are composed of a publicly known semiconductor memory and the like, and the above shooting image data is finally stored in the storage media 25 .
  • the shooting image data generated in the second embodiment conforms to the Exif (Exchangeable image file format for digital still cameras) standard, and a main body of the shooting image data and supplementary information (shooting information and so on) on the shooting image data are stored in association with each other.
  • the liquid crystal display 21 is connected to the monitor I/F 20 .
  • the liquid crystal display 21 is mainly placed at a rear portion of the electronic camera.
  • the viewfinder images sequentially outputted from the image processing section 16 are displayed by moving images on the liquid crystal display 21 during shooting.
  • a replay image plane of the shooting image data, a setting image plane to change various kinds of settings of the electronic camera, and so on are also displayed on the liquid crystal display 21 .
  • the manipulation section 22 includes an input button to perform switching between various kinds of shooting modes (such as a shooting mode and a replay mode) of the electronic camera and input settings, a release button, and so on.
  • shooting modes such as a shooting mode and a replay mode
  • the CPU 23 controls the operation of each section of the electronic camera according to a sequence program stored in a ROM not shown. For example, the CPU 23 performs an AE calculation, a calculation of a white balance gain, and so on based on the through image signals. The CPU 23 generates the supplementary information on the shooting image data based on the Exif standards when the shutter is released. Especially in the second embodiment, the CPU 23 has the following functions.
  • the CPU 23 performs publicly known face recognition processing on the through image signals to detect a face area of a person within the image shooting plane. Then, the CPU 23 generates face recognition information indicating the position of the face area within the shooing image plane. In the second embodiment, the CPU 23 also detects a vertical direction of the face based on the positional relationship among face parts (eyes, a nose, a mouth, and so on) at the time of face recognition.
  • Japanese Unexamined Patent Application Publication No. 8-63597 (1) a method of extracting a contour of a flesh-colored area based on color and detecting a face by the degree of matching with a face contour template which is prepared in advance, (2) a method of finding an eye candidate area and detecting a face by the degree of matching with an eye template, (3) a method of finding a feature quantity defined from a two-dimensional Fourier transform result of a face candidate area found by a face contour template and a two-dimensional Fourier transform result of a face template image including eyes, a nose, a mouth, and so on which is prepared in advance and detecting a face by subjecting the feature amount to threshold processing, and the like are disclosed.
  • the CPU 23 performs a contrast detection system AF calculation based on through image signals of a specified focus area located in the image shooting plane.
  • the CPU 23 selects the specified focus area from among plural focus areas (a group of focus areas) arranged regularly within the image shooting plane based on the face recognition information.
  • all of the focus areas located within a rectangular area which surrounds the contour of the face area compose the specified focus area.
  • the specified focus area is set to match the range of the above AF frame. Incidentally, the range of the specified focus area rarely perfectly matches the face area, whereby a surrounding portion adjacent to the face area is included in the specified focus area, which causes a high contrast in a contour portion of the face area (See FIG. 17 ).
  • the CPU 23 when the focusing position is not detected by the first AF calculation (the specified focus area including the face area), the CPU 23 changes the focus area (focus area which is likely to include the body of the subject) located under the face area to the specified focus area. Note that, this change of the specified focus area is set with reference to the vertical direction of the face detected using the face parts by the CPU 23 .
  • the contrast detection system AF calculation is performed based on a principle that there is a correlation between the degree of the blur and the contrast of the image, and the contrast of the image becomes maximum when focus is achieved. More specifically, the CPU 23 first extracts a high-frequency component in a predetermined band by a band-pass filter from the through-image signals corresponding to the specified focus area. The CPU 23 then generates a focus evaluation value regarding the subject image in the specified focus area by integrating an absolute value of the high-frequency component. This focus evaluation value is maximum when the contrast is maximum at a focusing position.
  • the CPU 23 moves the focusing lens in a predetermined direction and compares focus evaluation values before and after the movement. If the focus evaluation value after the movement is larger, the contrast is regarded as trending higher, and the CPU 23 moves the focusing lens further in the same direction and performs the same calculation. On the other hand, if the focus evaluation value after the movement is smaller, the contrast is trending lower, and the CPU 23 moves the focusing lens in an opposite direction and performs the same calculation.
  • the CPU 23 searches for a peak of the focus evaluation value (a focusing position). The above operation is generally called a hill-climbing operation. Incidentally, if the focusing position is not detected in the specified focus area, the CPU 23 outputs focusing failure information to the image processing section 16 .
  • Step S 1101 The CPU 23 allows the image pickup device 13 to generate the through image signal at predetermined intervals.
  • the image processing section 16 generates a viewfinder image based on the through image signals.
  • the CPU 23 displays the viewfinder image by the moving image on the liquid crystal display 21 . Accordingly, the user can frame the subject by the viewfinder image displayed on the liquid crystal monitor 21 .
  • Step S 1102 The CPU 23 determines whether the release button is half pressed. If the release button is half pressed (YES side), the CPU 23 goes to S 1103 . On the other hand, if no force is applied to the release button (NO side), the CPU 23 stands by until the release button is half pressed.
  • Step S 1104 The CPU 23 determines whether the face area is detected in S 1103 . If the face area is detected (YES side), the CPU 23 goes to S 1105 . On the other hand, if the face area is not detected (NO side), the CPU 23 goes to S 1108 .
  • Step S 1105 The CPU 23 sets the focus areas within a rectangular area which surrounds the contour of the face area to the specified focus area (See FIG. 17 ). Then, the CPU 23 performs the AF calculation by the hill-climbing operation based on the through image signals of the specified focus area. Incidentally in the AF calculation in S 1105 , the image processing section 16 combines and displays the AF frame with the face area in the viewfinder image (See FIG. 18 ).
  • Step S 1106 The CPU 23 determines whether the focusing position is detected in the specified focus area (S 1105 ). If the focusing position is detected (YES side), the CPU 23 goes to S 1109 . On the other hand, if the focusing position is not detected (NO side), the CPU 23 generates the focusing failure information and goes to S 1107 .
  • Step S 1107 In this case, the CPU 23 changes the focus area located under the face area to the specified focus area. Then, the CPU 23 performs again the AF calculation in the specified area after the change, and thereafter goes to S 1109 . Also when the focusing position is not detected by this second AF calculation, the CPU 23 generates the focusing failure information. Incidentally, in the AF calculation in S 1107 , the image processing section 16 gives the indication of focusing failure by the AF frame of the viewfinder image based on the first or second focusing failure information.
  • Step S 1108 Meanwhile, in this case, there is no person within the image shooting plane or the face of the person as the subject is not detected. Therefore, the CPU 23 selects the focus area in the normal operation and performs the AF calculation.
  • Step S 1109 the CPU 23 shoots the subject and generates the shooting image data by the user fully pressing the release button.
  • the CPU 23 records supplementary information such as the presence or absence of face recognition and the position of the specified focus area used for the AF calculation in the shooting image data.
  • Step S 1110 The CPU 23 determines whether there is a shooting ending instruction inputted by the user. If there is the shooting ending instruction (YES side), the CPU 23 stops the generation of the through image signals and so on, and ends the shooting. On the other hand, if there is no shooting ending instruction (NO side), the CPU 23 returns to S 1102 and repeats a series of operations. The above is the description of the shooting operation of the second embodiment.
  • the CPU 23 performs the AF calculation in the specified focus area including the contour of the face area, whereby the person in the image shooting plane can be easily brought into focus.
  • a high contrast occurs in the contour portion of the face area. Accordingly, compared with when the focus is detected only in a portion with a low contrast within the face area, a search for a contrast peak becomes easier in the second embodiment. Namely, focusing accuracy in the face of the subject increases.
  • the detection of the contour of the face area is relatively easy, which reduces a possibility that the focusing accuracy is influenced by the expression of the face of the subject.
  • the CPU 23 when the focusing position is not detected by the first AF calculation with the face area as the specified focus area, the CPU 23 performs the second AF calculation in the specified area where the body of the person is located (S 1107 ). Accordingly, even if focus cannot be achieved in the face area, the person as the subject can be brought into focus with a high probability.
  • the CPU 23 estimates the position of the body from the direction of the face and sets the second specified focus area. Accordingly, in the second AF calculation, the stable focusing accuracy can be ensured regardless of a shooting attitude the electronic camera such as the normal position or vertical position.
  • the AF frame is combined with the face area in the viewfinder image for display (S 1105 ).
  • the viewfinder image is displayed with the indication of focusing failure using the AF frame (S 1107 ).
  • the first and second focusing failure displays are different, so that the user can relatively easily judge whether the person is brought into focus from the display state of the AF frame.
  • the shooting image data contains the supplementary information such as the presence or absence of face recognition, the position of the specified focus area used for the AF calculation. Accordingly, referring to the supplementary information of the shooting image data with a viewer such as a personal computer, the user can know the situation at the time the shooting was made ex post facto.
  • FIG. 19 is a block diagram showing an overview of an electronic camera of a third embodiment.
  • the same numerals and symbols are used to designate components common to the second embodiment, and a description thereof will be omitted.
  • the third embodiment is a modified example of the second embodiment, and its configuration differs from that of the second embodiment in that an attitude sensor 26 is connected to the CPU 23 .
  • the attitude sensor 26 detects a shooting attitude in which the electronic camera is held in a normal position, an upper right vertical position shooting attitude in which the right side of the electronic camera is located at an upper position, an upper light vertical shooting attitude in which the left side of the electronic camera is located at an upper position, and an inverted position shooting attitude in which the electronic camera is inverted.
  • the CPU 23 changes the focus area located under the face area to the specified focus area based on an output of the attitude sensor 26 .
  • FIG. 20 is a flowchart showing a shooting operation in a fourth embodiment.
  • steps except S 1205 of the fourth embodiment correspond to steps of the second embodiments, respectively, and a duplicate description will be omitted.
  • a block diagram of an electronic camera in the fourth embodiment is common to the second embodiment or the third embodiment, and it will be not shown.
  • step S 1205 The CPU 23 sets, out of focus areas overlapping the contour of the face area, only part of the focus areas to the specified focus area based on the face recognition information (S 1203 ) (See FIG. 21 ). Then, the CPU 23 performs the AF calculation by the hill-climbing operation based on the through image signals of the specified focus area.
  • the contour portion of the face area is set to the specified focus area also in S 1205 , the focusing accuracy in the face of the subject can be increased similarly to the second embodiment.
  • the CPU 23 set the focus area overlapping the upper contour or the side contour of the face area to the specified focus area.
  • the specified focus area is selected based on the vertical direction of the face detected by the CPU 23 or the output of the attitude sensor 26 .
  • the specified focus area is smaller than that in the second embodiment, so that the calculation amount in the AF calculation is also reduced. Accordingly, the fourth embodiment makes it possible to simplify the circuit configuration of the CPU 23 and further speed up the AF calculation.
  • the second AF calculation may be performed, for example, in the focus area at the center of the image shooting plane, regardless of a result of the face recognition. Further, in the fourth embodiment, when there is a focusing failure in the first specified focus area, the second AF calculation may be performed in another focus area of the face area. Furthermore, in the second embodiment, the focus failure indication may be given only when there is a focusing failure in the second AF calculation.

Abstract

An electronic camera includes a face detecting section, a setting section, and a controlling section. The face detecting section detects a face of a subject. The setting section sets a scene shooting mode to adjust a shooting condition to an optimum shooting condition in accordance with each pre-assumed shooting scene. The controlling section controls the face detection of the face detecting section only when the setting section has set a scene shooting mode for shooting a scene including a person.

Description

    CROSS REFERENCE TO RELATED APPLICATION
  • This application is based upon and claims the benefit of priority from Japanese Patent Applications Nos. 2005-037675 and 2005-038424, both filed on Feb. 15, 2005, the entire contents of which are incorporated herein by reference.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention relates to an electronic camera which captures an image of a subject, and particularly relates to an electronic camera which can detecting a characteristic portion of a subject, for example, a face.
  • 2. Description of the Related Art
  • There has been proposed a camera which, in order to focus on a face of a subject when the subject is a person, detects the face portion of the person in a field captured by the camera and autofocuses on the detected face area.
  • However, it is conceivable that such problems that the face is not detectable or detection accuracy is low may occur depending on a shooting situation or shooting environment, which may lead to confusing a user.
  • For example, Japanese Unexamined Patent Application Publication No. 2001-215403 discloses an electronic camera having a face recognition function and performing focusing control according to the eyes of a subject.
  • However, the electronic camera in Japanese Unexamined Patent Application Publication No. 2001-215403 has a difficulty in focusing on the eyes of the subject when the subject closes his or her eyes or wears glasses, and the focusing operation thereof is low in stability. Therefore, there is still room for improvement.
  • SUMMARY OF THE INVENTION
  • The present invention is made to solve any one of the above problems in the related art. An object of the present invention is to provide an electronic camera which prevents a user from being confused. Another object of the present invention is to provide an electronic camera which can stably focus on a person as a subject by recognizing his/her face.
  • Hereinafter, the present invention will be described.
  • An electronic camera according to a first aspect of the present invention includes a face detecting section, a setting section, and a controlling section. The face detecting section detects a face of a subject. The setting section sets a scene shooting mode to adjust a shooting condition to an optimum shooting condition in accordance with each pre-assumed shooting scene. The controlling section controls the face detection of the face detecting section only when the setting section has set a scene shooting mode for shooting a scene including a person.
  • According to the electronic camera of the above first aspect, it is desirable that the scene shooting mode for shooting the scene including a person be a portrait shooting mode.
  • According to the electronic camera of the above first aspect, it is desirable that the controlling section does not allow the face detecting section to perform the face detection when the portrait shooting mode is a portrait shooting mode for shooting a night landscape.
  • According to the electronic camera of the above first aspect, it is desirable that the controlling section control a shooting lens to focus on a face area detected by the face detecting section.
  • According to the electronic camera of the above first aspect, it is desirable that it further include a function setting section which sets a function for each scene shooting mode, and that the scene shooting mode for shooting the scene including a person is provided with a setting item regarding the face detection.
  • According to the electronic camera of the above first aspect, it is desirable that it stop a digital zoom function of electronically magnifying a magnifying power during the face detection by the face detecting section.
  • According to the electronic camera of the above first aspect, it is desirable that it stop a closeup shooting function of shifting a shooting lens for closeup shooting during the face detection by the face detecting section.
  • According to the electronic camera of the above first aspect, it is desirable that it further include a display section which displays a subject image obtained before shooting, and that during the face detection by the face detecting section, the amount of shooting information for display on the display section is reduced compared with while the face detection is not performed.
  • An electronic camera according to a second aspect of the present invention includes an image pickup device, a face recognizing section, a focus area specifying section, and a focusing section. The image pickup device photoelectrically converts a subject image obtained by an optical shooting system to generate an image signal of an image shooting plane. The face recognizing section detects a face area in the image shooting plane according to the image signal. The focus area specifying section sets as a specified focus area, a focus area including a contour of the face area among a group of focus areas arranged in the image shooting plane. The focusing section calculates a focus evaluation value of the subject image according to the image signal corresponding to the specified focus area and detects as a focusing position a position of the optical shooting system when the focus evaluation value is maximum.
  • According to the electronic camera of the above second aspect, it is desirable that the focus area specifying section set, as the specified focus area, a part of plural focus areas including the contour of the face area. In this case, it is particularly desirable that the focus area specifying section set, as the specified focus area, a focus area overlapping with the contour of the face area at an upper side or a side thereof.
  • According to the electronic camera of the above second aspect, it is desirable that the focus area specifying section change the specified focus area to a focus area located below the face area, when the focusing position is not detected in the focus area including the contour of the face area.
  • According to the electronic camera of the above second aspect, the face recognizing section detects a direction of the face based on a positional relation of face parts in the face area. It is desirable that the focus area specifying section change a position of the focus area to be the specified focus area, according to the direction of the face.
  • According to the electronic camera of the above second aspect, it is desirable that it further include an attitude detecting section which detects a shooting attitude of the electronic camera, and that the focus area specifying section change a position of the focus area to be the specified focus area, according to the shooting attitude.
  • According to the electronic camera of the above second aspect, it is desirable that it further include an electronic viewfinder. The electronic viewfinder displays a viewfinder image of the image shooting plane according to the image signal, and displays an indication of focusing failure associated with a face area of the viewfinder image when the focusing position is not detected in the specified focus area.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The nature, principle, and utility of the invention will become more apparent from the following detailed description when read in conjunction with the accompanying drawings in which like parts are designated by identical reference numbers in which:
  • FIG. 1 is an external view of an electronic camera 1 according to a first embodiment;
  • FIG. 2 is a block diagram showing functions of the electronic camera 1 according to the first embodiment;
  • FIG. 3 is views showing a mode select dial 105 for selecting a shooting scene mode and a menu screen corresponding to the selected shooting scene mode;
  • FIG. 4 is views showing the mode select dial 105 for selecting a shooting scene mode and a menu screen corresponding to the selected shooting scene mode;
  • FIG. 5 is a flowchart showing control performed by a CPU 111;
  • FIG. 6 is a flowchart showing control performed by the CPU 111 in a face recognizing AF mode;
  • FIG. 7 is a flowchart showing face detection AF control performed by the CPU 111;
  • FIG. 8 is a flowchart showing AF control by a detection area performed by the CPU 111;
  • FIG. 9 is a flowchart showing AF control, in which a central area is weighted, performed by the CPU 111;
  • FIG. 10 is a view showing a display example of an image plane displayed on a monitor 103;
  • FIG. 11 is a view showing a display example of the image plane displayed on the monitor 103;
  • FIG. 12 is a view showing a display example of the image plane displayed on the monitor 103;
  • FIG. 13 is a view showing a display example of the image plane displayed on the monitor 103;
  • FIG. 14 is a view showing a display example of the image plane displayed on the monitor 103;
  • FIG. 15 is a block diagram showing an overview of an electronic camera of a second embodiment;
  • FIG. 16 is a flowchart showing a shooting operation in the second embodiment;
  • FIG. 17 is a view showing the position of a specified focus area in the second embodiment;
  • FIG. 18 is a view showing a viewfinder image at the time of face recognition in the second embodiment;
  • FIG. 19 is a block diagram showing an overview of an electronic camera of a third embodiment;
  • FIG. 20 is a flowchart showing a shooting operation in a fourth embodiment; and
  • FIG. 21 is a view showing the position of a specified focus area in the fourth embodiment.
  • DESCRIPTION OF THE PREFERRED EMBODIMENTS
  • Hereinafter, a description is given of embodiments of the invention with reference to the accompanying drawings.
  • Configuration of First Embodiment
  • First, the configuration of an electronic camera 1 according to a first embodiment of the present invention will be described.
  • FIG. 1 is an external view of the electronic camera 1 according to the first embodiment. In FIG. 1, the electronic camera 1 includes a release button 101, a cruciform key 102, a monitor 103, a decision button 104, a mode select dial 105, a zoom button, 106, a menu button 107, a play button 108, a closeup shooting button 109, and an optical viewfinder 110.
  • The release button 101 is a button capable of detecting two-stage operations: a half-press stage and a full-press stage. The release button 101 is manipulated by a user when the user instructs the start of shooting. The cruciform key 102 is manipulated by the user to move a cursor or the like on the monitor 103. The decision button 104 is a button manipulated by the user when the user selects and decides an item with the cruciform key 102 or the like. The decision button 104 is manipulated by the user also when the user switches on/off states of the monitor 103.
  • The mode select dial 105 is a dial which enables the user to change a camera function such as a shooting scene selection by turning it. The zoom button 106 is a button manipulated by the user when the user optically and electronically scales up or down an image recorded at the time of shooting. The zoom button 106 is manipulated by the user also when at the time of replay of image data, the user electronically scales up or down replayed image. The menu button 107 is a button manipulated by the user when the user wants to display a menu screen. The play button 108 is a button manipulated by the user when the user replays image data recorded in a memory. The closeup shooting button 109 is a button manipulated by the user when the user shoots a close subject such as a close plant. The optical viewfinder 110 is for the user to optically check a field.
  • FIG. 2 is a block diagram showing functions of the electronic camera 1 according to the first embodiment. In FIG. 2, the electronic camera 1 is composed of a CPU 111, a face detecting section 112, a built-in memory 113, a manipulation circuit 114, a display circuit 115, a monitor 103, an image processing circuit 116, an image pickup device 117, a zoom lens 118, a zoom driver 119, a focus lens 120, a focus driver 121, an aperture 122, an aperture driver 123, and a memory card slot 124. Needless to say, there are other circuits to realize functions of the electronic camera but have little relation to the first embodiment, so that a description thereof will be omitted.
  • The CPU 111 is a circuit which processes programs to realize various functions executed in the electronic camera 1. The CPU 111 executes the programs stored in a memory in the CPU 111, that is, the built-in memory 113 and controls various circuits in the electronic camera 1. The face detecting section 112 extracts a characteristic portion of image data picked up by the image pickup device 117 and detects a face area, face size, and so on of a subject. In FIG. 2, a function block called the face detecting section 112 is described for explanation, but in the first embodiment, a face detection function is realized in software by a face detection program executed by the CPU 111. Of course, it is possible to realize the face detecting section 112 by a hardware circuit.
  • The built-in memory 113 is a memory to store image data, a control program, and so on. For example, a nonvolatile semiconductor memory is used as the built-in memory 113. The built-in memory 113 stores the face detection program which is executed to detect the face area of the subject. Further, the built-in memory 113 can store face information such as the face position and face size obtained by face detection. The manipulation circuit 114 detects manipulations of manipulation buttons such as the release button 101, the cruciform key 102, and the decision button 104 provided in the electronic camera 1 and transfers them to the CPU 111. Further, the manipulation circuit 114 detects a half-press manipulation and a full-press manipulation of the release button 101. The display circuit 115 is a circuit to generate image plane data displayed on the monitor 103. The monitor 103 is a liquid crystal display provided on a rear surface of the electronic camera 1. This monitor 103 displays the image plane data generated by the display circuit 115.
  • A shooting lens is an optical lens to focus a subject image onto a light-receiving plane of the image pickup device 117. This shooting lens is composed of the zoom lens 118, the focus lens 120, and so on. Out of the lenses composing the shooting lens, the zoom lens 118 is a lens to realize scale-up and scale-down of the optical image focused on the image pickup device 117. This zoom lens 118 is moved by a motor. The zoom driver 119 is a circuit to drive the motor by a command of the CPU 111 and move the zoom lens 119 to a predetermined position. Out of the lenses composing the shooting lens, the focus lens 120 is a lens to adjust focus. This focus lens 120 is moved by a motor. The focus driver 121 is a circuit to drive the motor by a command of the CPU 111 and move the focus lens 120 to a predetermined position.
  • The aperture 122 is to adjust the amount of light of the field incident on the image pickup device 117. The aperture driver 123 is a circuit to drive a motor by a command of the CPU 111 and control open and closed states of the aperture 122. The image pickup device 117 is a device to convert the optical image inputted through the shooting lens into electric image signals. For example, the image pickup device 118 is composed of a CCD, or the like. The image processing circuit 116 analog-to-digital converts the electric signals outputted from the image pickup device 117 to generate digital signals. Further, the image processing circuit 116 performs interpolation processing or the like on the digital converted signals to generate image data. A memory card is inserted into the memory card slot 124. The memory card slot 124 writes data such as image data in the memory card or deletes data in the memory card.
  • Next, an operation in the electronic camera 1 according to the first embodiment will be described. A setting manipulation to use the face detection function in the electronic camera 1 will be described. When the main subject is a person, the user of the camera wants to focus on a face portion obtained by face detection. In the electronic camera 1, a result of face detection is used as one of options for deciding an area brought into autofocus (AF).
  • FIG. 3 and FIG. 4 are views each showing the mode select dial 105 to select a shooting scene mode and a menu screen corresponding to the selected shooting scene mode.
  • FIG. 3A shows the mode select dial 105 in a state where a portrait shooting mode is selected. FIG. 3B shows a menu screen displayed on the monitor 103 when the menu button 107 is manipulated in the portrait shooting mode. The user selects a function icon by manipulating the cruciform key 102 on the menu screen and manipulates the decision button 104. Thus, the electronic camera 1 performs a selected function. An icon 201 in FIG. 3B is an icon selected when a face detection AF function is performed. The operation of face detection AF will be described later.
  • FIG. 4A shows the mode select dial 105 in a state where a night landscape portrait shooting mode is selected. FIG. 4B shows a menu screen displayed on the monitor 103 when the menu button 107 is manipulated in the night landscape portrait shooting mode. As just shown, in the electronic camera 1 of the first embodiment, the menu screen of the night landscape portrait mode is not provided with the icon 201 which is selectable from the menu screen of the portrait mode.
  • A person is sometimes shot in the night landscape portrait mode. Therefore, it is effective to detect a face even in the night landscape portrait mode and autofocus on a detected face area. However, to detect the face of the subject, some degree of brightness of a face portion is required, but it is expected that the brightness is not sufficient to detect the face in a scene of the night landscape portrait. Hence, in order to prevent the user from being confused because the face detection is not feasible, the face detection AF is made unsettable in the night landscape portrait mode. Similarly, also when a sport shooting mode is selected via the mode select dial 105, it is difficult to detect a face of a moving person. Therefore, the sport shooting mode is not provided with the icon to select the face detection AF either. This results in avoiding in advance the user from being confused because the user cannot recognize the face.
  • When the face detection AF of the icon 201 on the menu screen shown in FIG. 3B is selected, the electronic camera 1 performs control so as to focus on the closest face portion of the detected face. When the face detection AF is selected, the electronic camera 1 is automatically switched to a constant AF mode. In the constant AF mode, irrespective of the manipulation of the release button 101, the electronic camera 1 repeats focusing by AF. Then, by half pressing the release button 101, the electronic camera 1 performs control to make an AF lock.
  • Subsequently, a face detection operation in the electronic camera 1 according to the first embodiment will be described.
  • In the electronic camera 1, the face detection is performed using a moving image (a through image) picked up by the image pickup device 117. This through image is displayed on the monitor 103 for check-up of a subject image to be shot. Moreover, during the face detection, the electronic camera 1 performs control in such a manner as to increase the brightness of the through image as compared with while the face detection is not performed. Increasing the brightness of the through image makes it easy to detect the face.
  • Further, the electronic camera 1 repeatedly performs the face detection using the through image until the release button 101 is half pressed. When the face is not detected at this time, the face detection is assisted by performing central-area AF and multi-area AF to bring the subject into focus.
  • Furthermore, the electronic camera 1 repeatedly stores face detection information such as the face position and face size as a result of the face detection in the built-in memory 113 while overwriting the latest piece thereof until the release button 101 is half pressed. This makes it possible to read and use the preceding detection result stored in the built-in memory 113 even if the face is not detected through the half-press manipulation. Besides, when the face detection information cannot be obtained in the half-press manipulation, the electronic camera 1 forcibly performs AF using another area such as a central area.
  • Hereinafter, shooting control by the face detection will be specifically described using a flowchart.
  • FIG. 5 is a flowchart showing control performed by the CPU 111. The flow shown in FIG. 5 is set to a still image shooting mode, and starts by detecting the manipulation of the menu button 107.
  • First, in step S01, the CPU 111 detects whether a scene shooting mode in which it is assumed that a person is shot is selected. In the electronic camera 1 of the first embodiment, it is detected whether the portrait mode or the night landscape portrait mode is selected by the mode select dial 105. If it is selected, the CPU 111 goes to step S02. On the other hand, if it is not selected, the CPU 111 goes to step S07. In step S02, the CPU 111 determines whether the selected shooting scene is the night landscape portrait mode. If it is the nigh landscape portrait mode, the CPU goes to step S07. On the other hand, if it is not the night landscape portrait mode, the CPU goes to step S03.
  • In step S03, the menu screen having the face detection AF as an option which is shown in FIG. 3B is displayed on he monitor 103. Subsequently, in step S04, the CPU 111 detects whether the decision button 104 is manipulated. If the manipulation of the decision button 104 is detected, the CPU 111 goes to step SOS. On the other hand, if the manipulation thereof is not detected, the CPU 111 continues the detection. In step 505, the CPU 111 determines whether the face detection AF is selected. If the face detection AF is selected, the CPU 111 goes to step S06. On the other hand, if any option other than the face detection AF is selected, the CPU 111 goes to step S09. In step S06, the CPU 111 controls the electronic camera 1 in a face recognition AF mode. The face recognition AF mode will be described using a flowchart in FIG. 6.
  • On the other hand, in step S07, the menu screen shown in FIG. 4B is displayed on the monitor 103. Subsequently, in step S08, the CPU 111 detects whether the decision button 104 is manipulated. If the manipulation thereof is detected, the CPU 111 goes to step S09. On the other hand, if the manipulation thereof is not detected, the CPU 111 continues the detection. Then, in step S09, the CPU 111 controls the electronic camera 1 in a normal AF mode.
  • Through the above control, the face detection is prevented from performed in the night landscape mode in which it is probable that the face cannot be detected, even during the scene shooting mode for shooting a person, which prevents the user from being confused.
  • Next, the face recognition AF mode will be described. FIG. 6 is the flowchart showing control performed by the CPU 111 in the face recognition AF mode. This flow is executed in step S06 in FIG. 5.
  • First, in step S51, the CPU 111 makes a setting not to perform an electronic zoom function. By preventing the electronic zoom function, the face detection can be continued using the through image. Then, in step S52, the CPU 111 makes a setting not to perform a closeup shooting function. This is because through the closeup shooting only part of the face is shot and the part of the face is not sufficient to detect the face. Subsequently, in step S53, the CPU 111 prohibits the monitor 103 from turning off. Without the monitor 103 being in an on state, the user cannot check up the result of face detection, so that he/she takes a shot even when the face detection of the electronic camera 1 is erroneous. Thereafter, in step S54, the CPU 111 switches the display of the monitor 103 to a simple display. In the simple display, the monitor 103 displays a reduced number of items of shooting information such as a memory remaining capacity indication which is superimposed on image data. A display example of shooting information in the face detection AF mode is shown in FIG. 11. A display example of shooting information in the normal AF mode is shown in FIG. 14. Switching to the simple display such as shown in FIG. 11 can make it easier for the user to check up a face detection frame as much as possible. This is because in the simple display, it is unlikely that the frame displayed when the face is detected overlaps with the display of the shooting information. Moreover, for example, even during the display setting to display the shooting information, the CPU 111 stops the shooting information display. When the shooting information display is not set, it is needless to say that the CPU 111 continues the non-display of the shooting information.
  • In step S55, the CPU 111 measures the brightness of the field using the image captured by the image pickup device 117. In step S56, the CPU 111 adjusts the brightness based on the measured field brightness. Then, the CPU 111 displays the through image on the monitor 103. In step S57, the CPU 111 performs face detection AF control.
  • The above control can reduce defects in the face detection AF mode and user's dissatisfaction.
  • Next, the face detection AF control will be described. FIG. 7 is a flowchart showing the face detection AF control performed by the CPU 111. This flow is executed in step S57 in FIG. 6.
  • First, in step S101, the CPU 111 determines whether the face is detected. If the face is detected, the CPU 111 goes to step S102. On the other hand, if the face is not detected, the CPU 111 goes to step S106. In step S102, if a face character shown in FIG. 10 is displayed while being superimposed on the subject image on the monitor 103, the CPU 111 deletes the display of the face character. Then, in step S103, the CPU 111 sets an area depending on the position and size of the detected face, and as shown in FIG. 11, displays frames while superimposing them on the subject image on the monitor 103. If the number of detected faces is plural, a frame is displayed on each face. In step S104, the CPU 111 performs AF in the set area. In step S105, the CPU 111 temporarily stores detected face information in the built-in memory 113. By storing the detected face information here, the CPU 111 can decide an AF area by using the face information stored in the built-in memory 113 when the face is not detected in the half-press manipulation in which the AF area is finally decided. The face information stored in the built-in memory 113 is overwritten with new face information in the next face detection.
  • On the other hand, in step S106, the CPU 111 displays the face character shown in FIG. 10 while superimposing it on the subject image on the monitor 103. The display of the face character indicates the user that the face detection AF mode is being performed, and informs the user of the size of the face ideal for the face detection control by the size of the face character. Incidentally, it is not necessary to display the face character constantly, and it is only required to display it every several seconds. Then, in step S107, the CPU 111 performs AF in an AF area in which a central area is weighted. There is a possibility that although the person is within the field, the face is too blurred to be detected, and hence, the face detection is assisted by focusing on the subject image in a central portion with a high possibility that the main subject is there to focus. In step S108, the CPU 111 determines again whether the face is detected. If the face is not detected, the CPU 111 goes to step S109. On the other hand, if the face is detected, the CPU goes to step S102. In step S109, the CPU 111 performs multi-area AF. Similarly to AF in the central area in step S107, a face of a person outside the central portion can be detected.
  • Next, in step S110, the CPU 111 determines again whether the face is detected. If the face is not detected, the CPU 111 goes to step S111. On the other hand, if the face is detected, the CPU 111 goes to step S102. In step S111, the CPU 111 detects whether the release button 101 is half pressed. If the half-press manipulation is detected, the CPU 111 goes to step S113. On the other hand, if the half-press manipulation is not detected, the CPU 111 goes to step S112. In step S112, the CPU 111 deletes the face information stored in the built-in memory 113 and returns to step S101.
  • In step S113, if the face character is displayed while being superimposed on the subject image on the monitor 103, the CPU 111 deletes the display of the face character. Then, in step S114, the CPU 111 determines whether the face is detected in order to specify the face which is regarded as a final AF area. If the face is detected, the CPU 111 goes to step S115. On the other hand, if the face is not detected, the CPU 111 goes to step S116. In step S115, the CPU 111 sets an area set according to the position and size of the detected face as the final AF area, and performs AF control. This AF control in the area detected by the face detection will be described later using FIG. 8. In step S116, the CPU 111 detects whether there is the face information stored in step S107 in the memory. With the face information therein, the CPU 111 goes to step S117. On the other hand, without the face information therein, the CPU 111 goes to step S118.
  • In step S117, the CPU 111 sets an area based on the stored face information as the AF area and performs AF control. The AF control in the area detected by the stored face detection will be described later using FIG. 8. Therefore, if the face is not detected in the half-press manipulation, the face area detected immediately therebefore with little time difference is used as the AF area. This makes it possible to focus on the face portion almost without fail. In particular, this can respond to one-press manipulation in which the release button 101 is fully-pressed with one press. In step S118, the CPU 111 performs AF control with the central area as the AF area. The AF control in the central area will be described later using FIG. 9. Consequently, when the face is not detected, the central area which is likely to include the main subject is automatically used as the AF area. This increases a possibility that the main subject comes into focus. Moreover, it becomes unnecessary for the user to return to the menu screen and reset the AF area, which prevents a photo opportunity from being missed.
  • In step S119, the CPU 111 detects whether the release button 101 is fully pressed. If the release button 101 is fully pressed, the CPU 111 goes to step S120. On the other hand, if it is not fully pressed, the CPU 111 goes to step S121. In step S120, the CPU 111 performs shooting and recording processing. In step S121, the CPU 111 detects whether the release button 101 is half pressed. If the release button 101 is half pressed, the CPU 111 returns to step S119. On the other hand, if it is not half pressed, the CPU 111 returns to step S101.
  • Next, the AF control in the detected area where the face has been detected will be described. FIG. 8 is a flowchart showing the AF control in the detected area performed by the CPU 111. This flow is executed in step S15 and step S117 in FIG. 7.
  • First, in step S201, the CPU 111 displays a face detected area frame set corresponding to the position and size of the detected face while superimposing it on the through image as shown in FIG. 12. When plural faces are detected, frames are displayed on the respective faces. Then, the frame of the largest or closest face is switched from a white frame (a thin-line frame in FIG. 11) to a red frame (a thick-line frame in FIG. 12). Subsequently, in step S202, the CPU 111 sets the area displayed by the red frame in FIG. 12 to the AF area and performs AF control. Thereafter, in step S203, the CPU 111 detects whether focus is achieved. If focus is achieved, the CPU 111 goes to step S204. On the other hand, if focus is not achieved, the CPU 111 goes to step S205. In step S204, the CPU 111 switches the frame shown by the thick-line frame in FIG. 12 from the red frame to a blue frame. This makes it possible to inform the user that focus is achieved. Moreover, as shown in FIG. 13, a frame display which is not to the AF area may be deleted. Meanwhile, in FIG. 205, the CPU 111 displays the frame shown by the thick-line frame in FIG. 12 which remains the red frame in a blinking state. This makes it possible to inform the user that focus is not achieved.
  • Next, the AF control in step S118 in FIG. 7 will be described. FIG. 9 is a flowchart showing the AF control in which the central area is weighted performed by the CPU 111. This flow is executed in step S118 in FIG. 7.
  • First, in step S301, the CPU 111 performs the AF control with the central area being weighted. Then, in step S302, the CPU 111 detects whether focus is achieved. If focus is achieved, the CPU 111 goes to step S303. On the other hand, if focus is not achieved, the CPU 111 goes to step S304. In step S303, the CPU 111 displays an icon 301 shown in FIG. 14 in an on-state while superimposing it on the subject image on the monitor 103. This makes it possible to inform the user that focus is achieved in an area other than the face detected area. On the other hand, in step S304, the CPU 111 displays the icon 301 shown in FIG. 14 in a blinking state while superimposing it on the subject image on the monitor 103. This makes it possible to inform the user that focus is not achieved in the area other than the face detected area.
  • Description of Second Embodiment
  • FIG. 15 is a block diagram showing an overview of an electronic camera of a second embodiment. The electronic camera of the second embodiment includes a shooting lens 11, lens driving mechanisms 12, an image pickup device 13, an analog signal processing section 14, an A/D conversion section 15, an image processing section 16, a compression/decompression section 17, a memory 18, a card I/F 19, a monitor I/F 20 and a liquid crystal display 21, a manipulation section 22, a CPU 23, and a bus 24. Incidentally, the image processing section 16, the compression/decompression section 17, the memory 18, the card I/F 19, the monitor I/F 20, and the CPU 23 are connected respectively via the bus 24.
  • The shooting lens 11 is composed of a group of plural lenses including a focusing lens for adjusting the focusing position. The position of this shooing lens 11 in an optical axis direction is adjusted by the lens driving mechanisms 12.
  • The image pickup device 13 is placed on the image space side of the shooting lens 11. Photodetectors which photoelectrically convert the subject image to generate analog image signals are two-dimensionally arranged on a light-receiving plane (a plane facing the shooting lens 11) of the image pickup device 13. An output of the image pickup device 13 is connected to the analog signal processing section 14.
  • Further, even when a shutter is not released, the image pickup device 13 exposes the subject at predetermined intervals and outputs the analog image signals (through image signals) by thinning-out reading or the like. This through image signal is used for AF calculation, AE calculation, and face recognition by the CPU 23, generation of a viewfinder moving image by the image processing section 26, and so on. Incidentally, the image pickup device 13 of the second embodiment may adopt either a sequential charge transfer method (for example, a CCD) or an XY address method (for example, a CMOS).
  • The analog signal processing section 14 is composed of a CDS circuit which performs correlated double sampling, a gain circuit which amplifies the outputs of the analog image signals, a clamp circuit which clamps the waveform of an input signal at a fixed voltage level, and so on. The A/D conversion section 15 converts the analog image signals outputted from the analog image processing section 14 into digital image signals. An output of the A/D conversion section 15 is connected to the image processing section 16 and the CPU 23, respectively.
  • The image processing section 16 performs image processing (defective pixel correction, gamma correction, interpolation, color conversion, edge enhancement, and so on) on the digital image signals when the shutter is released to generate shooting image data. Further, the image processing section 16 generates viewfinder images sequentially based on the digital image signals (through image signals) when the shutter is not released.
  • Furthermore, the image processing section 16 combines and displays a rectangular AF frame showing a face area as an AF target with the viewfinder image based on face recognition information described later (See FIG. 18). Besides, the image processing section 16 gives an indication of focusing failure to the viewfinder image using the above AF frame based on focusing failure information described later. Examples of this indication of focusing failure are a way of displaying the AF frame in a blinking state, a way of making a change to the color of the AF frame in a normal state, and so on. Incidentally, when the AF calculation is made twice and each ends in focusing failure as described later, the image processing section 16 gives different indications of focusing failure respectively for the first and second times.
  • The compression/decompression section 17 performs processing of compressing the shooting image data after image processing in a JPEG format and processing of decompressing and reconstructing the image data compressed in the JPEG format. The memory 18 is composed of an SDRAM or the like and has a capacity capable of recording image data corresponding to plural frames. Image data before and after the image processing by the image processing section 16 is temporarily stored in this memory 18.
  • A connector to connect storage media 25 is formed in the card I/F 19. The storage media 25 are composed of a publicly known semiconductor memory and the like, and the above shooting image data is finally stored in the storage media 25. Incidentally, the shooting image data generated in the second embodiment conforms to the Exif (Exchangeable image file format for digital still cameras) standard, and a main body of the shooting image data and supplementary information (shooting information and so on) on the shooting image data are stored in association with each other.
  • The liquid crystal display 21 is connected to the monitor I/F 20. The liquid crystal display 21 is mainly placed at a rear portion of the electronic camera. The viewfinder images sequentially outputted from the image processing section 16 are displayed by moving images on the liquid crystal display 21 during shooting. A replay image plane of the shooting image data, a setting image plane to change various kinds of settings of the electronic camera, and so on are also displayed on the liquid crystal display 21.
  • The manipulation section 22 includes an input button to perform switching between various kinds of shooting modes (such as a shooting mode and a replay mode) of the electronic camera and input settings, a release button, and so on.
  • The CPU 23 controls the operation of each section of the electronic camera according to a sequence program stored in a ROM not shown. For example, the CPU 23 performs an AE calculation, a calculation of a white balance gain, and so on based on the through image signals. The CPU 23 generates the supplementary information on the shooting image data based on the Exif standards when the shutter is released. Especially in the second embodiment, the CPU 23 has the following functions.
  • First, the CPU 23 performs publicly known face recognition processing on the through image signals to detect a face area of a person within the image shooting plane. Then, the CPU 23 generates face recognition information indicating the position of the face area within the shooing image plane. In the second embodiment, the CPU 23 also detects a vertical direction of the face based on the positional relationship among face parts (eyes, a nose, a mouth, and so on) at the time of face recognition.
  • Incidentally, as an example of the face recognition processing, in Japanese Unexamined Patent Application Publication No. 8-63597, (1) a method of extracting a contour of a flesh-colored area based on color and detecting a face by the degree of matching with a face contour template which is prepared in advance, (2) a method of finding an eye candidate area and detecting a face by the degree of matching with an eye template, (3) a method of finding a feature quantity defined from a two-dimensional Fourier transform result of a face candidate area found by a face contour template and a two-dimensional Fourier transform result of a face template image including eyes, a nose, a mouth, and so on which is prepared in advance and detecting a face by subjecting the feature amount to threshold processing, and the like are disclosed.
  • Secondly, the CPU 23 performs a contrast detection system AF calculation based on through image signals of a specified focus area located in the image shooting plane. Here, the CPU 23 selects the specified focus area from among plural focus areas (a group of focus areas) arranged regularly within the image shooting plane based on the face recognition information. In the second embodiment, all of the focus areas located within a rectangular area which surrounds the contour of the face area compose the specified focus area. In the second embodiment, the specified focus area is set to match the range of the above AF frame. Incidentally, the range of the specified focus area rarely perfectly matches the face area, whereby a surrounding portion adjacent to the face area is included in the specified focus area, which causes a high contrast in a contour portion of the face area (See FIG. 17).
  • Moreover, in the second embodiment, when the focusing position is not detected by the first AF calculation (the specified focus area including the face area), the CPU 23 changes the focus area (focus area which is likely to include the body of the subject) located under the face area to the specified focus area. Note that, this change of the specified focus area is set with reference to the vertical direction of the face detected using the face parts by the CPU 23.
  • Here, the contrast detection system AF calculation is performed based on a principle that there is a correlation between the degree of the blur and the contrast of the image, and the contrast of the image becomes maximum when focus is achieved. More specifically, the CPU 23 first extracts a high-frequency component in a predetermined band by a band-pass filter from the through-image signals corresponding to the specified focus area. The CPU 23 then generates a focus evaluation value regarding the subject image in the specified focus area by integrating an absolute value of the high-frequency component. This focus evaluation value is maximum when the contrast is maximum at a focusing position.
  • Thereafter, the CPU 23 moves the focusing lens in a predetermined direction and compares focus evaluation values before and after the movement. If the focus evaluation value after the movement is larger, the contrast is regarded as trending higher, and the CPU 23 moves the focusing lens further in the same direction and performs the same calculation. On the other hand, if the focus evaluation value after the movement is smaller, the contrast is trending lower, and the CPU 23 moves the focusing lens in an opposite direction and performs the same calculation. By repeating the above processing, the CPU 23 searches for a peak of the focus evaluation value (a focusing position). The above operation is generally called a hill-climbing operation. Incidentally, if the focusing position is not detected in the specified focus area, the CPU 23 outputs focusing failure information to the image processing section 16.
  • The shooting operation in the second embodiment will be described below with reference to a flowchart in FIG. 16.
  • Step S1101: The CPU 23 allows the image pickup device 13 to generate the through image signal at predetermined intervals. The image processing section 16 generates a viewfinder image based on the through image signals. The CPU 23 displays the viewfinder image by the moving image on the liquid crystal display 21. Accordingly, the user can frame the subject by the viewfinder image displayed on the liquid crystal monitor 21.
  • Step S1102: The CPU 23 determines whether the release button is half pressed. If the release button is half pressed (YES side), the CPU 23 goes to S1103. On the other hand, if no force is applied to the release button (NO side), the CPU 23 stands by until the release button is half pressed.
  • Step S1103: The CPU 23 detects the face area of the subject within the image shooting plane based on the through image signals. Then, the CPU 23 generates the face recognition information when there is a face area within the image shooting plane.
  • Step S1104: The CPU 23 determines whether the face area is detected in S1103. If the face area is detected (YES side), the CPU 23 goes to S1105. On the other hand, if the face area is not detected (NO side), the CPU 23 goes to S1108.
  • Step S1105: The CPU 23 sets the focus areas within a rectangular area which surrounds the contour of the face area to the specified focus area (See FIG. 17). Then, the CPU 23 performs the AF calculation by the hill-climbing operation based on the through image signals of the specified focus area. Incidentally in the AF calculation in S1105, the image processing section 16 combines and displays the AF frame with the face area in the viewfinder image (See FIG. 18).
  • Step S1106: The CPU 23 determines whether the focusing position is detected in the specified focus area (S1105). If the focusing position is detected (YES side), the CPU 23 goes to S1109. On the other hand, if the focusing position is not detected (NO side), the CPU 23 generates the focusing failure information and goes to S1107.
  • Step S1107: In this case, the CPU 23 changes the focus area located under the face area to the specified focus area. Then, the CPU 23 performs again the AF calculation in the specified area after the change, and thereafter goes to S1109. Also when the focusing position is not detected by this second AF calculation, the CPU 23 generates the focusing failure information. Incidentally, in the AF calculation in S1107, the image processing section 16 gives the indication of focusing failure by the AF frame of the viewfinder image based on the first or second focusing failure information.
  • Step S1108: Meanwhile, in this case, there is no person within the image shooting plane or the face of the person as the subject is not detected. Therefore, the CPU 23 selects the focus area in the normal operation and performs the AF calculation.
  • Step S1109: Then, the CPU 23 shoots the subject and generates the shooting image data by the user fully pressing the release button. Incidentally, by using a MakerNote tag of the Exif standard when the shooing image data is generated, the CPU 23 records supplementary information such as the presence or absence of face recognition and the position of the specified focus area used for the AF calculation in the shooting image data.
  • Step S1110: The CPU 23 determines whether there is a shooting ending instruction inputted by the user. If there is the shooting ending instruction (YES side), the CPU 23 stops the generation of the through image signals and so on, and ends the shooting. On the other hand, if there is no shooting ending instruction (NO side), the CPU 23 returns to S1102 and repeats a series of operations. The above is the description of the shooting operation of the second embodiment.
  • Next, effects of the above second embodiment will be described.
  • (1) In the second embodiment, the CPU 23 performs the AF calculation in the specified focus area including the contour of the face area, whereby the person in the image shooting plane can be easily brought into focus. Especially in the specified focus area, a high contrast occurs in the contour portion of the face area. Accordingly, compared with when the focus is detected only in a portion with a low contrast within the face area, a search for a contrast peak becomes easier in the second embodiment. Namely, focusing accuracy in the face of the subject increases. Moreover, the detection of the contour of the face area is relatively easy, which reduces a possibility that the focusing accuracy is influenced by the expression of the face of the subject.
  • (2) In the second embodiment, when the focusing position is not detected by the first AF calculation with the face area as the specified focus area, the CPU 23 performs the second AF calculation in the specified area where the body of the person is located (S1107). Accordingly, even if focus cannot be achieved in the face area, the person as the subject can be brought into focus with a high probability. The CPU 23 estimates the position of the body from the direction of the face and sets the second specified focus area. Accordingly, in the second AF calculation, the stable focusing accuracy can be ensured regardless of a shooting attitude the electronic camera such as the normal position or vertical position.
  • (3) In the second embodiment, for the AF calculation the AF frame is combined with the face area in the viewfinder image for display (S1105). Hence, the user can easily keep track of the person as an AF target from the viewfinder image on the liquid crystal display 21. Further, in the second AF calculation, the viewfinder image is displayed with the indication of focusing failure using the AF frame (S1107). The first and second focusing failure displays are different, so that the user can relatively easily judge whether the person is brought into focus from the display state of the AF frame.
  • (4) In the second embodiment, the shooting image data contains the supplementary information such as the presence or absence of face recognition, the position of the specified focus area used for the AF calculation. Accordingly, referring to the supplementary information of the shooting image data with a viewer such as a personal computer, the user can know the situation at the time the shooting was made ex post facto.
  • Description of Third Embodiment
  • FIG. 19 is a block diagram showing an overview of an electronic camera of a third embodiment. In the description of the following embodiments, the same numerals and symbols are used to designate components common to the second embodiment, and a description thereof will be omitted.
  • The third embodiment is a modified example of the second embodiment, and its configuration differs from that of the second embodiment in that an attitude sensor 26 is connected to the CPU 23. The attitude sensor 26 detects a shooting attitude in which the electronic camera is held in a normal position, an upper right vertical position shooting attitude in which the right side of the electronic camera is located at an upper position, an upper light vertical shooting attitude in which the left side of the electronic camera is located at an upper position, and an inverted position shooting attitude in which the electronic camera is inverted. When the focusing position is not detected in the specified focus area including the face area, the CPU 23 changes the focus area located under the face area to the specified focus area based on an output of the attitude sensor 26.
  • In the third embodiment, almost the same effects as in the first embodiment can be obtained. Furthermore, since the position of the specified focus area is changed by the output of the attitude sensor 26, it is possible to reduce the calculation load of the CPU 23 regarding the detection of the face in the vertical direction.
  • Description of Fourth Embodiment
  • FIG. 20 is a flowchart showing a shooting operation in a fourth embodiment. Here, steps except S1205 of the fourth embodiment correspond to steps of the second embodiments, respectively, and a duplicate description will be omitted. Further, a block diagram of an electronic camera in the fourth embodiment is common to the second embodiment or the third embodiment, and it will be not shown.
  • step S1205: The CPU 23 sets, out of focus areas overlapping the contour of the face area, only part of the focus areas to the specified focus area based on the face recognition information (S1203) (See FIG. 21). Then, the CPU 23 performs the AF calculation by the hill-climbing operation based on the through image signals of the specified focus area.
  • Since the contour portion of the face area is set to the specified focus area also in S1205, the focusing accuracy in the face of the subject can be increased similarly to the second embodiment. However, in the focus area overlapping the lower contour of the face area, there is a high possibility that the contrast is lowered by the flesh-colored potion of a neck. Therefore, it is desirable that in S1205, the CPU 23 set the focus area overlapping the upper contour or the side contour of the face area to the specified focus area. Incidentally, when the upper or side focus area of the face area is used, the specified focus area is selected based on the vertical direction of the face detected by the CPU 23 or the output of the attitude sensor 26.
  • In this fourth embodiment, the specified focus area is smaller than that in the second embodiment, so that the calculation amount in the AF calculation is also reduced. Accordingly, the fourth embodiment makes it possible to simplify the circuit configuration of the CPU 23 and further speed up the AF calculation.
  • Supplementary Description of Embodiments
  • When there is a focusing failure in the specified focus area including the face area in the above second embodiment, the second AF calculation may be performed, for example, in the focus area at the center of the image shooting plane, regardless of a result of the face recognition. Further, in the fourth embodiment, when there is a focusing failure in the first specified focus area, the second AF calculation may be performed in another focus area of the face area. Furthermore, in the second embodiment, the focus failure indication may be given only when there is a focusing failure in the second AF calculation.
  • The invention is not limited to the above embodiments and various modifications may be made without departing from the spirit and scope of the invention. Any improvement may be made in part or all of the components.

Claims (18)

1. An electronic camera, comprising:
a face detecting section which detects a face of a subject;
a setting section which sets a scene shooting mode to adjust a shooting condition to an optimum shooting condition in accordance with each of pre-assumed shooting scenes; and
a controlling section which controls the face detection of said face detecting section only when said setting section has set a scene shooting mode for shooting a scene including a person.
2. The electronic camera according to claim 1, wherein
the scene shooting mode for shooting the scene including a person is a portrait shooting mode.
3. The electronic camera according to claim 2, wherein
said controlling section does not allow said face detecting section to perform the face detection when the portrait shooting mode is a portrait shooting mode for shooting a night landscape.
4. The electronic camera according to claim 1, wherein
said controlling section controls a shooting lens to focus on a face area detected by said face detecting section.
5. The electronic camera according to claim 1, further comprising
a function setting section which sets a function for each of scene shooting modes, wherein
the scene shooting mode for shooting the scene including a person is provided with a setting item regarding the face detection.
6. The electronic camera according to claim 1, wherein
during the face detection by said face detecting section, the electronic camera stops a digital zoom function of electronically magnifying a magnifying power.
7. The electronic camera according to claim 1, wherein
during the face detection by said face detecting section, the electronic camera stops a closeup shooting function of shifting a shooting lens for closeup shooting.
8. The electronic camera according to claim 1, further comprising
a display section which displays a subject image obtained before shooting, wherein
during the face detection by said face detecting section, an amount of shooting information for display on said display section is reduced compared with while the face detection is not performed.
9. An electronic camera, comprising:
an image pickup device which photoelectrically converts a subject image obtained by an optical shooting system to generate an image signal of an image shooting plane;
a face recognizing section which detects a face area in the image shooting plane according to said image signal;
a focus area specifying section which sets, as a specified focus area, a focus area including a contour of the face area among a group of focus areas arranged in said image shooting plane; and
a focusing section which calculates a focus evaluation value of said subject image according to the image signal corresponding to the specified focus area and detects a position of said optical shooting system as a focusing position when the focus evaluation value is maximum.
10. The electronic camera according to claim 9, wherein
said focus area specifying section sets, as the specified focus area, a part of plural focus areas including the contour of the face area.
11. The electronic camera according to claim 10, wherein
said focus area specifying section sets, as the specified focus area, a focus area overlapping with the contour of the face area at an upper side or a side thereof.
12. The electronic camera according to claim 9, wherein
said focus area specifying section changes the specified focus area to a focus area located below the face area, when the focusing position is not detected in the focus area including the contour of the face area.
13. The electronic camera according to claim 10, wherein
said focus area specifying section changes the specified focus area to a focus area located below the face area, when the focusing position is not detected in the focus area including the contour of the face area.
14. The electronic camera according to claim 11, wherein:
said face recognizing section detects a direction of the face based on a positional relation of face parts in said face area; and
said focus area specifying section changes a position of the focus area to be the specified focus area, according to the direction of the face.
15. The electronic camera according to claim 12, wherein:
said face recognizing section detects a direction of the face based on a positional relation of face parts in the face area; and
said focus area specifying section changes a position of the focus area to be the specified focus area, according to the direction of the face.
16. The electronic camera according to claim 11, further comprising
an attitude detecting section which detects a shooting attitude of the electronic camera, wherein
said focus area specifying section changes a position of the focus area to be the specified focus area, according to the shooting attitude.
17. The electronic camera according to claim 12, further comprising
an attitude detecting section which detects a shooting attitude of said electronic camera, wherein
said focus area specifying section changes a position of the focus area to be the specified focus area, according to the shooting attitude.
18. The electronic camera according to claim 9, further comprising
an electronic viewfinder which displays a viewfinder image of the image shooting plane according to the image signal, and displays an indication of focusing failure associated with a face area of the viewfinder image when the focusing position is not detected in the specified focus area.
US11/345,393 2005-02-15 2006-02-02 Electronic camera Abandoned US20060182433A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/289,747 US7881601B2 (en) 2005-02-15 2008-11-03 Electronic camera

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
JP2005-038424 2005-02-15
JP2005037675A JP4639837B2 (en) 2005-02-15 2005-02-15 Electronic camera
JP2005-037675 2005-02-15
JP2005038424A JP4581730B2 (en) 2005-02-15 2005-02-15 Digital camera

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US12/289,747 Continuation US7881601B2 (en) 2005-02-15 2008-11-03 Electronic camera

Publications (1)

Publication Number Publication Date
US20060182433A1 true US20060182433A1 (en) 2006-08-17

Family

ID=36815725

Family Applications (2)

Application Number Title Priority Date Filing Date
US11/345,393 Abandoned US20060182433A1 (en) 2005-02-15 2006-02-02 Electronic camera
US12/289,747 Expired - Fee Related US7881601B2 (en) 2005-02-15 2008-11-03 Electronic camera

Family Applications After (1)

Application Number Title Priority Date Filing Date
US12/289,747 Expired - Fee Related US7881601B2 (en) 2005-02-15 2008-11-03 Electronic camera

Country Status (1)

Country Link
US (2) US20060182433A1 (en)

Cited By (54)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070047941A1 (en) * 2005-08-31 2007-03-01 Nikon Corporation Autofocus apparatus
US20070065137A1 (en) * 2005-09-21 2007-03-22 Sony Corporation Photographic device, method of processing information, and program
US20070223912A1 (en) * 2006-03-27 2007-09-27 Fujifilm Corporation Photographing method and photographing apparatus
US20070242942A1 (en) * 2006-04-14 2007-10-18 Nikon Corporation Camera
US20070263997A1 (en) * 2006-05-10 2007-11-15 Canon Kabushiki Kaisha Focus adjustment method, focus adjustment apparatus, and control method thereof
US20070268397A1 (en) * 2006-05-12 2007-11-22 Fujifilm Corporation Image pickup apparatus and image pickup control method
US20070286568A1 (en) * 2006-05-19 2007-12-13 Casio Computer Co., Ltd. Scene selection screen generation device, scene selection screen adding system, method for generating scene selection screen, method for adding a scene selection screen, imaging device, and computer readable storage medium having recorded thereon a program
US20080043135A1 (en) * 2006-08-15 2008-02-21 Fujifilm Corporation Photographing apparatus and in-focus position searching method
US20080074529A1 (en) * 2006-09-22 2008-03-27 Sony Corporation Imaging apparatus, control method of imaging apparatus, and computer program
US20080170129A1 (en) * 2007-01-17 2008-07-17 Samsung Techwin Co., Ltd. Digital photographing apparatus and method for controlling the same
US20080170132A1 (en) * 2007-01-17 2008-07-17 Samsung Techwin Co., Ltd. Digital photographing apparatus, method for controlling the same, and a recording medium for storing a program to implement the method
US20080193116A1 (en) * 2007-02-09 2008-08-14 Canon Kabushiki Kaisha Focusing device and image-capturing device provided with the same
US20080239086A1 (en) * 2007-03-28 2008-10-02 Fujifilm Corporation Digital camera, digital camera control process, and storage medium storing control program
US20080266439A1 (en) * 2007-04-26 2008-10-30 Sony Corporation Image pickup apparatus
US20080278589A1 (en) * 2007-05-11 2008-11-13 Karl Ola Thorn Methods for identifying a target subject to automatically focus a digital camera and related systems, and computer program products
US20080309785A1 (en) * 2007-06-14 2008-12-18 Masahiko Sugimoto Photographing apparatus
US20090002516A1 (en) * 2007-06-28 2009-01-01 Sony Corporation Image capturing apparatus, shooting control method, and program
US20090009622A1 (en) * 2007-07-03 2009-01-08 Canon Kabushiki Kaisha Image data management apparatus and method, and recording medium
WO2009008541A1 (en) * 2007-07-10 2009-01-15 Canon Kabushiki Kaisha Focus control appratus, image sensing apparatus, and control method therefor
US20090059058A1 (en) * 2007-08-31 2009-03-05 Yuuki Okabe Image pickup apparatus and focusing condition displaying method
US20090059059A1 (en) * 2007-08-27 2009-03-05 Sanyo Electric Co., Ltd. Electronic Camera
US20090207298A1 (en) * 2008-02-19 2009-08-20 Canon Kabushiki Kaisha Focusing apparatus and method for controlling the same
US20090251590A1 (en) * 2008-04-07 2009-10-08 Canon Kabushiki Kaisha Image capturing apparatus and control method therefor
US20090284645A1 (en) * 2006-09-04 2009-11-19 Nikon Corporation Camera
US20090322934A1 (en) * 2008-06-30 2009-12-31 Canon Kabushiki Kaisha Focus detection apparatus and control method therefor
US20090322933A1 (en) * 2008-06-30 2009-12-31 Canon Kabushiki Kaisha Image capturing apparatus and control method for the same
EP2141537A1 (en) * 2008-07-04 2010-01-06 Canon Kabushiki Kaisha Image pickup apparatus, method of controlling the same, computer program code, and storage medium
US20100002127A1 (en) * 2008-07-04 2010-01-07 Canon Kabushiki Kaisha Image pickup apparatus and auto-focus detection method
US20100079589A1 (en) * 2008-09-26 2010-04-01 Sanyo Electric Co., Ltd. Imaging Apparatus And Mode Appropriateness Evaluating Method
US20100097515A1 (en) * 2008-10-22 2010-04-22 Canon Kabushiki Kaisha Auto focusing apparatus and auto focusing method, and image sensing apparatus
US20100097513A1 (en) * 2008-10-21 2010-04-22 Sony Corporation Image pickup apparatus and method for controlling the same
US20100171836A1 (en) * 2009-01-07 2010-07-08 Canon Kabushiki Kaisha Image capturing apparatus, control method thereof, and program
US20100208127A1 (en) * 2009-02-12 2010-08-19 Sony Corporation Image capturing apparatus, control method thereof, and program
US20100295962A1 (en) * 2009-05-20 2010-11-25 Hoya Corporation Imaging apparatus and hdri method
US7869701B1 (en) * 2007-09-21 2011-01-11 Tessera MEMS Technologies, Inc. Adaptive autofocus lens positioning
EP2309721A1 (en) * 2006-10-12 2011-04-13 Nikon Corporation Digital camera
US20110085057A1 (en) * 2008-07-01 2011-04-14 Nikon Corporation Imaging device, image display device, and electronic camera
US20110249142A1 (en) * 2010-04-07 2011-10-13 Apple Inc. Face Detection Using Orientation Sensor Data
US20110273471A1 (en) * 2009-01-19 2011-11-10 Sony Corporation Display control device, display control method and program
US8068164B2 (en) 2007-09-14 2011-11-29 Sony Corporation Face recognition auto focus apparatus for a moving image
US20110305446A1 (en) * 2010-06-15 2011-12-15 Kei Itoh Imaging apparatus, focus position detecting method, and computer program product
US8289439B2 (en) 2008-04-09 2012-10-16 Canon Kabushiki Kaisha Image capturing apparatus and control method therefor
US20130194479A1 (en) * 2010-09-17 2013-08-01 Fujifilm Corporation Display control method and device for finder device
EP1956831B1 (en) * 2007-02-08 2013-10-09 Canon Kabushiki Kaisha Focus adjusting device, image pickup apparatus, and focus adjustment method
US20140035952A1 (en) * 2011-04-20 2014-02-06 Nec Casio Mobile Communications, Ltd. Individual identification character display system, terminal device, individual identification character display method, and computer program
CN103731604A (en) * 2012-10-15 2014-04-16 奥林巴斯映像株式会社 Tracking apparatus and tracking method
US20150249782A1 (en) * 2012-09-21 2015-09-03 Canon Kabushiki Kaisha Image processing apparatus, image processing method, program, and image pickup apparatus having the image processing apparatus
US20160044233A1 (en) * 2008-11-14 2016-02-11 Canon Kabushiki Kaisha Image capturing apparatus capable of storing focus detection data
US20160330366A1 (en) * 2014-01-17 2016-11-10 Sony Corporation Image processing device and method, and program
US20160353026A1 (en) * 2015-05-29 2016-12-01 Thomson Licensing Method and apparatus for displaying a light field based image on a user's device, and corresponding computer program product
EP1988706B1 (en) * 2007-05-02 2017-11-08 Casio Computer Co., Ltd. Imaging device, imaging control program, and imaging control method
JP2018097380A (en) * 2018-02-01 2018-06-21 キヤノン株式会社 Imaging device and control method of the same
CN112422806A (en) * 2019-08-22 2021-02-26 三赢科技(深圳)有限公司 Camera device and corresponding automatic focusing system and method
US11050923B2 (en) * 2019-01-18 2021-06-29 Canon Kabushiki Kaisha Imaging apparatus and control method

Families Citing this family (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8068160B2 (en) * 2006-02-20 2011-11-29 Panasonic Corporation Imaging apparatus, printing apparatus, printing system, printing method, display system, display apparatus, and display method
JP4345757B2 (en) * 2006-02-22 2009-10-14 セイコーエプソン株式会社 Image data color correction
US7796163B2 (en) * 2006-07-25 2010-09-14 Fujifilm Corporation System for and method of taking image based on objective body in a taken image
JP4819001B2 (en) * 2006-07-25 2011-11-16 富士フイルム株式会社 Imaging apparatus and method, program, image processing apparatus and method, and program
JP5129638B2 (en) * 2008-04-02 2013-01-30 ペンタックスリコーイメージング株式会社 Imaging device
JP2010011441A (en) * 2008-05-26 2010-01-14 Sanyo Electric Co Ltd Imaging apparatus and image playback device
US8570429B2 (en) * 2009-02-27 2013-10-29 Samsung Electronics Co., Ltd. Image processing method and apparatus and digital photographing apparatus using the same
JP5257157B2 (en) 2009-03-11 2013-08-07 ソニー株式会社 IMAGING DEVICE, IMAGING DEVICE CONTROL METHOD, AND PROGRAM
US8717490B2 (en) * 2009-06-19 2014-05-06 Casio Computer Co., Ltd Imaging apparatus, focusing method, and computer-readable recording medium recording program
JP2012103979A (en) * 2010-11-11 2012-05-31 Sanyo Electric Co Ltd Image processing apparatus
US8811747B2 (en) * 2011-10-28 2014-08-19 Intellectual Ventures Fund 83 Llc Image recomposition from face detection and facial features
US9025835B2 (en) 2011-10-28 2015-05-05 Intellectual Ventures Fund 83 Llc Image recomposition from face detection and facial features
US8938100B2 (en) 2011-10-28 2015-01-20 Intellectual Ventures Fund 83 Llc Image recomposition from face detection and facial features
US9025836B2 (en) 2011-10-28 2015-05-05 Intellectual Ventures Fund 83 Llc Image recomposition from face detection and facial features
US9008436B2 (en) 2011-10-28 2015-04-14 Intellectual Ventures Fund 83 Llc Image recomposition from face detection and facial features
JP5970937B2 (en) * 2012-04-25 2016-08-17 ソニー株式会社 Display control apparatus and display control method
JP6238581B2 (en) * 2013-06-10 2017-11-29 キヤノン株式会社 Imaging apparatus and focus detection method
KR102423364B1 (en) * 2015-08-12 2022-07-22 삼성전자 주식회사 Method for providing image and electronic device supporting the same

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6301440B1 (en) * 2000-04-13 2001-10-09 International Business Machines Corp. System and method for automatically setting image acquisition controls
US20030071908A1 (en) * 2001-09-18 2003-04-17 Masato Sannoh Image pickup device, automatic focusing method, automatic exposure method, electronic flash control method and computer program
US20040119852A1 (en) * 2002-12-12 2004-06-24 Lg Electronics Inc. Zoom method of a digital camera
US20040208114A1 (en) * 2003-01-17 2004-10-21 Shihong Lao Image pickup device, image pickup device program and image pickup method
US20050046730A1 (en) * 2003-08-25 2005-03-03 Fuji Photo Film Co., Ltd. Digital camera
US20070065134A1 (en) * 2005-09-16 2007-03-22 Fuji Photo Film Co., Ltd. Target-image position detecting apparatus, method and program for controlling said apparatus
US20070248345A1 (en) * 2006-04-04 2007-10-25 Nikon Corporation Camera

Family Cites Families (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2592985B2 (en) 1990-09-25 1997-03-19 キヤノン株式会社 Imaging device
JPH07270671A (en) 1994-03-31 1995-10-20 Nikon Corp Automatic focusing camera
JP2001215403A (en) 2000-02-01 2001-08-10 Canon Inc Image pickup device and automatic focus detection method
JP2001304855A (en) 2000-04-18 2001-10-31 Olympus Optical Co Ltd Range finder
JP4536248B2 (en) 2000-11-24 2010-09-01 オリンパス株式会社 Imaging device
JP2003107335A (en) 2001-09-28 2003-04-09 Ricoh Co Ltd Image pickup device, automatic focusing method, and program for making computer execute the method
JP3927802B2 (en) 2001-12-19 2007-06-13 株式会社リコー Image processing device
JP2003344891A (en) 2002-05-23 2003-12-03 Canon Inc Automatic photographing mode setting camera
JP2004023656A (en) 2002-06-19 2004-01-22 Canon Inc Image processing device, image processing method, and program
JP2004117776A (en) 2002-09-26 2004-04-15 Fuji Photo Film Co Ltd Digital camera
JP4126721B2 (en) 2002-12-06 2008-07-30 富士フイルム株式会社 Face area extraction method and apparatus
US20040207743A1 (en) 2003-04-15 2004-10-21 Nikon Corporation Digital camera system
JP4127521B2 (en) * 2003-05-20 2008-07-30 富士フイルム株式会社 Digital camera and control method thereof
JP2005215750A (en) 2004-01-27 2005-08-11 Canon Inc Face detecting device and face detecting method

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6301440B1 (en) * 2000-04-13 2001-10-09 International Business Machines Corp. System and method for automatically setting image acquisition controls
US20030071908A1 (en) * 2001-09-18 2003-04-17 Masato Sannoh Image pickup device, automatic focusing method, automatic exposure method, electronic flash control method and computer program
US20040119852A1 (en) * 2002-12-12 2004-06-24 Lg Electronics Inc. Zoom method of a digital camera
US20040208114A1 (en) * 2003-01-17 2004-10-21 Shihong Lao Image pickup device, image pickup device program and image pickup method
US20050046730A1 (en) * 2003-08-25 2005-03-03 Fuji Photo Film Co., Ltd. Digital camera
US20070065134A1 (en) * 2005-09-16 2007-03-22 Fuji Photo Film Co., Ltd. Target-image position detecting apparatus, method and program for controlling said apparatus
US20070248345A1 (en) * 2006-04-04 2007-10-25 Nikon Corporation Camera

Cited By (111)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070047941A1 (en) * 2005-08-31 2007-03-01 Nikon Corporation Autofocus apparatus
US7660519B2 (en) * 2005-08-31 2010-02-09 Nikon Corporation Autofocus apparatus
US20070065137A1 (en) * 2005-09-21 2007-03-22 Sony Corporation Photographic device, method of processing information, and program
US7860389B2 (en) * 2005-09-21 2010-12-28 Sony Corporation Photographic device, method of processing information, and program
US20070223912A1 (en) * 2006-03-27 2007-09-27 Fujifilm Corporation Photographing method and photographing apparatus
US8073319B2 (en) * 2006-03-27 2011-12-06 Fujifilm Corporation Photographing method and photographing apparatus based on face detection and photography conditions
US20070242942A1 (en) * 2006-04-14 2007-10-18 Nikon Corporation Camera
US7706675B2 (en) * 2006-04-14 2010-04-27 Nikon Corporation Camera
US7945152B2 (en) 2006-05-10 2011-05-17 Canon Kabushiki Kaisha Focus adjustment method, focus adjustment apparatus, and control method thereof
US20100226636A1 (en) * 2006-05-10 2010-09-09 Canon Kabushiki Kaisha Focus adjustment method, focus adjustment apparatus, and control method thereof
US8145049B2 (en) 2006-05-10 2012-03-27 Canon Kabushiki Kaisha Focus adjustment method, focus adjustment apparatus, and control method thereof
US20070263997A1 (en) * 2006-05-10 2007-11-15 Canon Kabushiki Kaisha Focus adjustment method, focus adjustment apparatus, and control method thereof
CN101547314B (en) * 2006-05-10 2013-01-16 佳能株式会社 Focus adjustment apparatus and control method thereof
US20070268397A1 (en) * 2006-05-12 2007-11-22 Fujifilm Corporation Image pickup apparatus and image pickup control method
US20070286568A1 (en) * 2006-05-19 2007-12-13 Casio Computer Co., Ltd. Scene selection screen generation device, scene selection screen adding system, method for generating scene selection screen, method for adding a scene selection screen, imaging device, and computer readable storage medium having recorded thereon a program
US7909700B2 (en) * 2006-05-19 2011-03-22 Casio Computer Co., Ltd. Scene selection screen generation device, scene selection screen adding system, method for generating scene selection screen, method for adding a scene selection screen, imaging device, and computer readable storage medium having recorded thereon a program
US20080043135A1 (en) * 2006-08-15 2008-02-21 Fujifilm Corporation Photographing apparatus and in-focus position searching method
US8310586B2 (en) * 2006-08-15 2012-11-13 Fujifilm Corporation Photographing apparatus and in-focus position searching method
US20090284645A1 (en) * 2006-09-04 2009-11-19 Nikon Corporation Camera
US8538252B2 (en) * 2006-09-04 2013-09-17 Nikon Corporation Camera
US20080074529A1 (en) * 2006-09-22 2008-03-27 Sony Corporation Imaging apparatus, control method of imaging apparatus, and computer program
US7929042B2 (en) * 2006-09-22 2011-04-19 Sony Corporation Imaging apparatus, control method of imaging apparatus, and computer program
EP2309721A1 (en) * 2006-10-12 2011-04-13 Nikon Corporation Digital camera
US20080170132A1 (en) * 2007-01-17 2008-07-17 Samsung Techwin Co., Ltd. Digital photographing apparatus, method for controlling the same, and a recording medium for storing a program to implement the method
CN101242494B (en) * 2007-01-17 2012-03-21 三星电子株式会社 Digital photographing apparatus and method for controlling the same
KR101323735B1 (en) * 2007-01-17 2013-10-30 삼성전자주식회사 Digital photographing apparatus, method for controlling the same, and recording medium storing program to implement the method
US20080170129A1 (en) * 2007-01-17 2008-07-17 Samsung Techwin Co., Ltd. Digital photographing apparatus and method for controlling the same
US8368764B2 (en) * 2007-01-17 2013-02-05 Samsung Electronics Co., Ltd. Digital photographing apparatus and method for controlling the same
US8063943B2 (en) * 2007-01-17 2011-11-22 Samsung Electronics Co., Ltd. Digital photographing apparatus, method for controlling the same, and a recording medium for storing a program to implement the method
EP1956831B1 (en) * 2007-02-08 2013-10-09 Canon Kabushiki Kaisha Focus adjusting device, image pickup apparatus, and focus adjustment method
US20080193116A1 (en) * 2007-02-09 2008-08-14 Canon Kabushiki Kaisha Focusing device and image-capturing device provided with the same
US7747159B2 (en) 2007-02-09 2010-06-29 Canon Kabushiki Kaisha Focusing device and image-capturing device provided with the same
US20080239086A1 (en) * 2007-03-28 2008-10-02 Fujifilm Corporation Digital camera, digital camera control process, and storage medium storing control program
US7907202B2 (en) * 2007-04-26 2011-03-15 Sony Corporation Image pickup apparatus
US20080266439A1 (en) * 2007-04-26 2008-10-30 Sony Corporation Image pickup apparatus
EP1988706B1 (en) * 2007-05-02 2017-11-08 Casio Computer Co., Ltd. Imaging device, imaging control program, and imaging control method
WO2008138409A1 (en) * 2007-05-11 2008-11-20 Sony Ericsson Mobile Communications Ab Methods for identifying a target subject to automatically focus a digital camera and related systems, and computer program products
US20080278589A1 (en) * 2007-05-11 2008-11-13 Karl Ola Thorn Methods for identifying a target subject to automatically focus a digital camera and related systems, and computer program products
CN102438099A (en) * 2007-06-14 2012-05-02 富士胶片株式会社 Digital image pickup apparatus
CN102438105A (en) * 2007-06-14 2012-05-02 富士胶片株式会社 Photographing apparatus
US20080309785A1 (en) * 2007-06-14 2008-12-18 Masahiko Sugimoto Photographing apparatus
US20110234851A1 (en) * 2007-06-14 2011-09-29 Masahiko Sugimoto Photographing apparatus
EP2015568A3 (en) * 2007-06-14 2012-02-22 FUJIFILM Corporation Digital image pickup apparatus
US9065998B2 (en) 2007-06-14 2015-06-23 Fujifilm Corporation Photographing apparatus provided with an object detection function
US9131138B2 (en) 2007-06-14 2015-09-08 Fujifilm Corporation Photographing apparatus
US20090002516A1 (en) * 2007-06-28 2009-01-01 Sony Corporation Image capturing apparatus, shooting control method, and program
US8823864B2 (en) * 2007-06-28 2014-09-02 Sony Corporation Image capturing apparatus and associated methodology for auto-focus and facial detection
US20090009622A1 (en) * 2007-07-03 2009-01-08 Canon Kabushiki Kaisha Image data management apparatus and method, and recording medium
US8466977B2 (en) * 2007-07-03 2013-06-18 Canon Kabushiki Kaisha Image data management apparatus and method, and recording medium
US8279324B2 (en) * 2007-07-10 2012-10-02 Canon Kabushiki Kaisha Focus control apparatus, image sensing apparatus, and control method therefor
US20100013981A1 (en) * 2007-07-10 2010-01-21 Canon Kabushiki Kaisha Focus control apparatus, image sensing apparatus, and control method therefor
WO2009008541A1 (en) * 2007-07-10 2009-01-15 Canon Kabushiki Kaisha Focus control appratus, image sensing apparatus, and control method therefor
CN102360150A (en) * 2007-07-10 2012-02-22 佳能株式会社 Focus control apparatus, image sensing apparatus, and control method therefor
US8077252B2 (en) * 2007-08-27 2011-12-13 Sanyo Electric Co., Ltd. Electronic camera that adjusts a distance from an optical lens to an imaging surface so as to search the focal point
US20090059059A1 (en) * 2007-08-27 2009-03-05 Sanyo Electric Co., Ltd. Electronic Camera
US20090059058A1 (en) * 2007-08-31 2009-03-05 Yuuki Okabe Image pickup apparatus and focusing condition displaying method
US8106998B2 (en) 2007-08-31 2012-01-31 Fujifilm Corporation Image pickup apparatus and focusing condition displaying method
US8068164B2 (en) 2007-09-14 2011-11-29 Sony Corporation Face recognition auto focus apparatus for a moving image
US7869701B1 (en) * 2007-09-21 2011-01-11 Tessera MEMS Technologies, Inc. Adaptive autofocus lens positioning
US8502912B2 (en) 2008-02-19 2013-08-06 Canon Kabushiki Kaisha Focusing apparatus and method for controlling the same
US20090207298A1 (en) * 2008-02-19 2009-08-20 Canon Kabushiki Kaisha Focusing apparatus and method for controlling the same
US20090251590A1 (en) * 2008-04-07 2009-10-08 Canon Kabushiki Kaisha Image capturing apparatus and control method therefor
US8026975B2 (en) 2008-04-07 2011-09-27 Canon Kabushiki Kaisha Image capturing apparatus and control method therefor
US8289439B2 (en) 2008-04-09 2012-10-16 Canon Kabushiki Kaisha Image capturing apparatus and control method therefor
US8471951B2 (en) * 2008-04-09 2013-06-25 Canon Kabushiki Kaisha Image capturing apparatus and control method therefor
US20090322934A1 (en) * 2008-06-30 2009-12-31 Canon Kabushiki Kaisha Focus detection apparatus and control method therefor
US8248516B2 (en) 2008-06-30 2012-08-21 Canon Kabushiki Kaisha Focus detection apparatus and control method therefor
US8279323B2 (en) * 2008-06-30 2012-10-02 Canon Kabushiki Kaisha Image capturing apparatus and control method for the same
US20090322933A1 (en) * 2008-06-30 2009-12-31 Canon Kabushiki Kaisha Image capturing apparatus and control method for the same
US20110085057A1 (en) * 2008-07-01 2011-04-14 Nikon Corporation Imaging device, image display device, and electronic camera
US20150341588A1 (en) * 2008-07-01 2015-11-26 Nikon Corporation Imaging device, image display device, and electronic camera that determines whether to record the position at which an image is photographed and the accuracy of the photographic position to be recorded
EP2141537A1 (en) * 2008-07-04 2010-01-06 Canon Kabushiki Kaisha Image pickup apparatus, method of controlling the same, computer program code, and storage medium
US20100002128A1 (en) * 2008-07-04 2010-01-07 Canon Kabushiki Kaisha Image pickup apparatus, method of controlling the same, and storage medium
US20100002127A1 (en) * 2008-07-04 2010-01-07 Canon Kabushiki Kaisha Image pickup apparatus and auto-focus detection method
US9188837B2 (en) * 2008-07-04 2015-11-17 Canon Kabushiki Kaisha Image pickup apparatus, method of controlling the same, and storage medium
US8259214B2 (en) 2008-07-04 2012-09-04 Canon Kabushiki Kaisha Image pickup apparatus and auto-focus detection method
US20100079589A1 (en) * 2008-09-26 2010-04-01 Sanyo Electric Co., Ltd. Imaging Apparatus And Mode Appropriateness Evaluating Method
US20100097513A1 (en) * 2008-10-21 2010-04-22 Sony Corporation Image pickup apparatus and method for controlling the same
US8305481B2 (en) * 2008-10-21 2012-11-06 Sony Corporation Image pickup apparatus and method for controlling the same
US20100097515A1 (en) * 2008-10-22 2010-04-22 Canon Kabushiki Kaisha Auto focusing apparatus and auto focusing method, and image sensing apparatus
US8330849B2 (en) * 2008-10-22 2012-12-11 Canon Kabushiki Kaisha Auto focusing apparatus and auto focusing method, and image sensing apparatus for stable focus detection
US20160044233A1 (en) * 2008-11-14 2016-02-11 Canon Kabushiki Kaisha Image capturing apparatus capable of storing focus detection data
US8477194B2 (en) 2009-01-07 2013-07-02 Canon Kabushiki Kaisha Image capturing apparatus, control method thereof, and program
US20100171836A1 (en) * 2009-01-07 2010-07-08 Canon Kabushiki Kaisha Image capturing apparatus, control method thereof, and program
US20110273471A1 (en) * 2009-01-19 2011-11-10 Sony Corporation Display control device, display control method and program
US8390730B2 (en) 2009-02-12 2013-03-05 Sony Corporation Image capturing apparatus, control method thereof, and program
US20100208127A1 (en) * 2009-02-12 2010-08-19 Sony Corporation Image capturing apparatus, control method thereof, and program
US20100295962A1 (en) * 2009-05-20 2010-11-25 Hoya Corporation Imaging apparatus and hdri method
US8305453B2 (en) * 2009-05-20 2012-11-06 Pentax Ricoh Imaging Company, Ltd. Imaging apparatus and HDRI method
US8405736B2 (en) * 2010-04-07 2013-03-26 Apple Inc. Face detection using orientation sensor data
US20110249142A1 (en) * 2010-04-07 2011-10-13 Apple Inc. Face Detection Using Orientation Sensor Data
US8724981B2 (en) * 2010-06-15 2014-05-13 Ricoh Company, Limited Imaging apparatus, focus position detecting method, and computer program product
US20110305446A1 (en) * 2010-06-15 2011-12-15 Kei Itoh Imaging apparatus, focus position detecting method, and computer program product
US20130194479A1 (en) * 2010-09-17 2013-08-01 Fujifilm Corporation Display control method and device for finder device
US8872957B2 (en) * 2010-09-17 2014-10-28 Fujifilm Corporation Display control method and device for finder device
US9721388B2 (en) * 2011-04-20 2017-08-01 Nec Corporation Individual identification character display system, terminal device, individual identification character display method, and computer program
US20140035952A1 (en) * 2011-04-20 2014-02-06 Nec Casio Mobile Communications, Ltd. Individual identification character display system, terminal device, individual identification character display method, and computer program
US9826139B2 (en) * 2012-09-21 2017-11-21 Canon Kabushiki Kaisha Image processing apparatus, image processing method, program, and image pickup apparatus having the image processing apparatus
US20150249782A1 (en) * 2012-09-21 2015-09-03 Canon Kabushiki Kaisha Image processing apparatus, image processing method, program, and image pickup apparatus having the image processing apparatus
US9761010B2 (en) 2012-10-15 2017-09-12 Olympus Corporation Tracking apparatus
US9317748B2 (en) * 2012-10-15 2016-04-19 Olympus Corporation Tracking apparatus
CN103731604A (en) * 2012-10-15 2014-04-16 奥林巴斯映像株式会社 Tracking apparatus and tracking method
US20140105454A1 (en) * 2012-10-15 2014-04-17 Olympus Imaging Corp. Tracking apparatus
US20160330366A1 (en) * 2014-01-17 2016-11-10 Sony Corporation Image processing device and method, and program
US10237466B2 (en) * 2014-01-17 2019-03-19 Sony Corporation Recognition of degree of focus of an image
US20160353026A1 (en) * 2015-05-29 2016-12-01 Thomson Licensing Method and apparatus for displaying a light field based image on a user's device, and corresponding computer program product
US10116867B2 (en) * 2015-05-29 2018-10-30 Thomson Licensing Method and apparatus for displaying a light field based image on a user's device, and corresponding computer program product
JP2018097380A (en) * 2018-02-01 2018-06-21 キヤノン株式会社 Imaging device and control method of the same
US11050923B2 (en) * 2019-01-18 2021-06-29 Canon Kabushiki Kaisha Imaging apparatus and control method
CN112422806A (en) * 2019-08-22 2021-02-26 三赢科技(深圳)有限公司 Camera device and corresponding automatic focusing system and method
US11115572B2 (en) * 2019-08-22 2021-09-07 Triple Win Technology (Shenzhen) Co. Ltd. Automatic focusing system, method, and vehicular camera device therefor

Also Published As

Publication number Publication date
US20090147107A1 (en) 2009-06-11
US7881601B2 (en) 2011-02-01

Similar Documents

Publication Publication Date Title
US7881601B2 (en) Electronic camera
US7791668B2 (en) Digital camera
JP4761146B2 (en) Imaging apparatus and program thereof
JP4457358B2 (en) Display method of face detection frame, display method of character information, and imaging apparatus
KR101342477B1 (en) Imaging apparatus and imaging method for taking moving image
JP4639837B2 (en) Electronic camera
US7706674B2 (en) Device and method for controlling flash
US8587658B2 (en) Imaging device, image display device, and program with intruding object detection
JP5025532B2 (en) Imaging apparatus, imaging apparatus control method, and imaging apparatus control program
US7668451B2 (en) System for and method of taking image
US20080181460A1 (en) Imaging apparatus and imaging method
US20090213239A1 (en) Imaging device and method for its image processing
US20090059054A1 (en) Apparatus, method, and recording medium containing program for photographing
JP4462305B2 (en) Imaging apparatus, method, and program
JP4657960B2 (en) Imaging method and apparatus
US9131138B2 (en) Photographing apparatus
WO2010073619A1 (en) Image capture device
JP5782813B2 (en) Imaging apparatus and image display method
US9148577B2 (en) Interchangeable-lens camera, and viewfinder display method
JP2007310813A (en) Image retrieving device and camera
US7885527B2 (en) Focusing apparatus and method
JP4645413B2 (en) Imaging device
JP4792929B2 (en) Digital camera
JP5045538B2 (en) Imaging apparatus and program
JP2009038749A (en) Image processor, control method of image processor, and program for implementing the same

Legal Events

Date Code Title Description
AS Assignment

Owner name: NIKON CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KAWAHARA, TAKUMI;MOTOKI, YASUYUKI;OHMURA, AKIRA;REEL/FRAME:017471/0165;SIGNING DATES FROM 20060319 TO 20060324

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION