US8587691B2 - Photographing apparatus and method for dynamic range adjustment and stereography - Google Patents

Photographing apparatus and method for dynamic range adjustment and stereography Download PDF

Info

Publication number
US8587691B2
US8587691B2 US12/625,776 US62577609A US8587691B2 US 8587691 B2 US8587691 B2 US 8587691B2 US 62577609 A US62577609 A US 62577609A US 8587691 B2 US8587691 B2 US 8587691B2
Authority
US
United States
Prior art keywords
image
imaging device
exposure
electric signal
photographing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US12/625,776
Other versions
US20100134652A1 (en
Inventor
Yasuo Takane
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Original Assignee
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Samsung Electronics Co Ltd filed Critical Samsung Electronics Co Ltd
Assigned to SAMSUNG DIGITAL IMAGING CO., LTD. reassignment SAMSUNG DIGITAL IMAGING CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: TAKANE, YASUO
Publication of US20100134652A1 publication Critical patent/US20100134652A1/en
Assigned to SAMSUNG ELECTRONICS CO., LTD. reassignment SAMSUNG ELECTRONICS CO., LTD. MERGER (SEE DOCUMENT FOR DETAILS). Assignors: SAMSUNG DIGITAL IMAGING CO., LTD.
Priority to US14/054,077 priority Critical patent/US8947557B2/en
Application granted granted Critical
Publication of US8587691B2 publication Critical patent/US8587691B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/239Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/70Circuitry for compensating brightness variation in the scene
    • H04N23/75Circuitry for compensating brightness variation in the scene by influencing optical camera components
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/70Circuitry for compensating brightness variation in the scene
    • H04N23/73Circuitry for compensating brightness variation in the scene by influencing the exposure time
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/45Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from two or more image sensors being of different type or operating in different modes, e.g. with a CMOS sensor for moving images in combination with a charge-coupled device [CCD] for still images
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/70Circuitry for compensating brightness variation in the scene
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/70Circuitry for compensating brightness variation in the scene
    • H04N23/71Circuitry for evaluating the brightness variation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/70Circuitry for compensating brightness variation in the scene
    • H04N23/741Circuitry for compensating brightness variation in the scene by increasing the dynamic range of the image compared to the dynamic range of the electronic image sensors

Definitions

  • the present invention relates to a photographing apparatus and method for dynamic range adjustment and stereography.
  • the subject When photographing of the subject is performed using a photographing apparatus having a narrow allowable dynamic range, the subject may not be photographed well. At this time, since the photographing apparatus is not in an appropriate exposure state, the captured image may include dark or bright smeared portions.
  • the dynamic range may be enlarged using the following operations.
  • the dynamic range may be enlarged by (1) non-linear treatment such as gamma or knee operation, (2) using a non-linear photographing apparatus, (3) combining images obtained from imaging devices having properties different from each other, or (4) image signal processing or combining images obtained from one imaging device.
  • each of the images may be captured after changing exposure, and thus, brightness of an electronic view finder (EVF) or a liquid crystal display (LCD) in the live view display may be repeatedly changed between dark and bright states, or continuity of displaying the subject in the live view display may not be ensured.
  • EMF electronic view finder
  • LCD liquid crystal display
  • the present invention provides a photographing apparatus and method, in which an exposure amount required to successfully perform a main photographing operation for a still image is ensured while displaying a moving picture image stably before the main photographing operation for the still image.
  • a photographing apparatus including: a first imaging device for converting light from a subject received through a first optical system into an electric signal; a second imaging device for converting light from the subject received through a second optical system into an electric signal; a first image signal processor for generating an image signal for a live view based on the electric signal output from the first imaging device before a photographing operation of a still image; an exposure controller for controlling an exposure so as to perform a step exposure in the second imaging device before the photographing of the still image; and an exposure calculator for calculating an exposure amount during the photographing operation of the still image based on the electric signal converted in the second imaging device obtained through the step exposure.
  • the photographing apparatus may further include: an exposure amount setting unit for setting an exposure amount of the first or second imaging device based on highlight reference and an exposure amount of the other imaging devices based on dark reference at the same time based on the calculated exposure amount; and a second image processor for generating a first image signal for a still image based on the electric signal converted by the first imaging device and a second image signal for a still image based on the electric signal converted by the second imaging device at the same time when the photographing operation of the still image is performed based on the set exposure amount.
  • an exposure amount setting unit for setting an exposure amount of the first or second imaging device based on highlight reference and an exposure amount of the other imaging devices based on dark reference at the same time based on the calculated exposure amount
  • a second image processor for generating a first image signal for a still image based on the electric signal converted by the first imaging device and a second image signal for a still image based on the electric signal converted by the second imaging device at the same time when the photographing operation of the still image is performed based
  • the photographing apparatus may further include: a dislocation detector for detecting a dislocation between the first and second image signals; and a combining unit for combining the first and second image signals with each other using the detected dislocation.
  • the photographing apparatus may further include: an exposure amount setting unit for setting exposure amounts of the first and second imaging devices to be the same as each other at the same timing based on the calculated exposure amount; and a second image processor for generating a first image signal for a still image based on the electric signal converted by the first imaging device and a second image signal for a still image based on the electric signal converted by the second imaging device at the same time when the photographing operation of the still image is performed based on the set exposure amount.
  • an exposure amount setting unit for setting exposure amounts of the first and second imaging devices to be the same as each other at the same timing based on the calculated exposure amount
  • a second image processor for generating a first image signal for a still image based on the electric signal converted by the first imaging device and a second image signal for a still image based on the electric signal converted by the second imaging device at the same time when the photographing operation of the still image is performed based on the set exposure amount.
  • the photographing apparatus may further include: a dislocation detector for detecting a dislocation between the first and second image signals; and a combining unit for combining the first and second image signals with each other using the detected dislocation.
  • a photographing method including: converting light from a subject received through a first optical system into an electric signal by a first imaging device; converting light from the subject received through a second optical system into an electric signal by a second imaging device; generating an image signal for a live view based on the electric signal converted by the first imaging device before a photographing operation of a still image; controlling an exposure to be a step exposure in the second imaging device before the photographing operation of the still image; and calculating an exposure amount required in the photographing operation of the still image based on the electric signal converted by the second imaging device, which is obtained by the step exposure.
  • FIG. 1 is a block diagram of a photographing apparatus according to an embodiment of the present invention.
  • FIG. 2 is a block diagram of a digital signal processor (DSP) in the photographing apparatus of FIG. 1 , according to an embodiment of the present invention
  • FIGS. 3A and 3B are flowcharts illustrating operations of the photographing apparatus of FIG. 1 in a dynamic range enlarging mode, according to an embodiment of the present invention
  • FIGS. 4A and 4B are flowcharts illustrating operations of the photographing apparatus of FIG. 1 in a stereo image photographing mode, according to an embodiment of the present invention.
  • FIG. 5 is a block diagram of a photographing apparatus according to another embodiment of the present invention.
  • a photographing apparatus 100 according to an embodiment of the present invention will be described as follows.
  • FIG. 1 is a block diagram of the photographing apparatus 100 according to the embodiment of the present invention.
  • the photographing apparatus 100 includes two photographing systems each of which including an optical system and an imaging system, and each of the photographing systems may separately photograph a subject to obtain an image. For example, when a live view image is displayed, an image displayed as the live view image may be obtained through one photographing system, and the other photographing system may measure light. Otherwise, the photographing systems may be arranged parallel to each other to photograph the subject at the same time, and thus, a stereo image (a three-dimensional image) may be generated.
  • the photographing apparatus 100 includes optical systems 101 and 102 , image sensors 103 and 104 , synchronous random access memories (SDRAMs) 105 and 106 , a memory card 107 , a read only memory 109 , a release button 115 , a timing generator 117 , digital signal processors 121 and 122 , a flash controller 131 , a flash 133 , a liquid crystal display (LCD) controller 135 , and an LCD 137 .
  • SDRAMs synchronous random access memories
  • LCD liquid crystal display
  • the optical systems 101 and 102 are respectively examples of a first optical system and a second optical system.
  • the optical systems 101 and 102 form image of external light on the image sensors 103 and 104 , respectively transmitting light from the subject toward the image sensors 103 and 104 .
  • the light from the optical system 101 reaches the image sensor 103
  • the light from the optical system 102 reaches the image sensor 104 .
  • Each of the optical systems 101 and 102 may include a zoom lens, a diaphragm, and a focusing lens.
  • the zoom lens changes a viewing angle by changing a focal length, and the diaphragm adjusts an amount of the light transmitting therethrough.
  • the focusing lens focuses an image of the subject on an imaging surface of the image sensor 103 or 104 by moving from a side to another side, or vice versa.
  • the image sensors 103 and 104 are examples of a first imaging device and a second imaging device (photoelectric conversion device), and may include photoelectric conversion devices that convert optical information incident from the optical systems 101 and 102 into electric signals.
  • the image sensors 103 and 104 may generate the electric signals according to the received light.
  • Charge coupled device (CCD) sensors or complementary metal oxide semiconductor (CMOS) sensors may be used as the image sensors 103 and 104 .
  • CCD Charge coupled device
  • CMOS complementary metal oxide semiconductor
  • a mechanical shutter that may block light when photographing is not performed and may transmit light when photographing is performed may be used to control exposure times of the image sensors 103 and 104 .
  • an electronic shutter (not shown) may be used.
  • the mechanical shutter or the electronic shutter may be operated by a switch connected to the DSPs 121 and 122 , such as the release button 115 (a manipulation member).
  • Each of the image sensors 103 and 104 respectively further includes a correlated double sampler (CDS)/amplifier (AMP), and an analog/digital (A/D) converter.
  • CDS correlated double sampler
  • AMP amplifier
  • A/D analog/digital converter
  • the CDS/AMP removes low-frequency noise included in the electric signal output from the image sensors 103 and 104 , and at the same time, each may amplify a respective electric signal to a predetermined level.
  • the A/D converter converts the electric signal output from the CDS/AMP into a digital signal.
  • the A/D converter outputs the digital signal to an image signal processor 141 .
  • the SDRAMs 105 and 106 temporarily store image data of captured images.
  • the SDRAMs 105 and 106 may store image data of a plurality of images.
  • the SDRAMs 105 and 106 are respectively connected to the DSPs 121 and 122 . Reading/writing of image data from/to the SDRAMs 105 and 106 are controlled by a memory controller 151 .
  • the SDRAM 105 may include a region as video RAM (VRAM).
  • VRAM video RAM
  • the VRAM may be used for displaying images, and may include a plurality of channels.
  • the VRAMs may simultaneously perform operations of inputting image data therein and outputting image data to the LCD controller 135 .
  • a definition or a maximum color representation number may depend on a capacity of the VRAM.
  • the memory card 107 may include a semiconductor memory medium such as a flash memory. Image data generated by a photographing operation is recorded in the memory card 107 or is read from the memory card 107 .
  • the memory card 107 is not limited to the flash memory, but may include an optical disc (a compact disc (CD), a digital versatile disc (DVD), or a blue-ray disc), a photomagnetic disc, or a magnetic disc.
  • the memory card 107 may be detachable from the photographing apparatus 100 .
  • the ROM 109 stores an operating program for the DSPs 121 and 122 .
  • the release button 115 may operate in a half-pushed state, a fully-pushed state, and a released state.
  • the release button 115 outputs a manipulation signal for starting focusing control when the release button 115 is half-pushed (S 1 operation).
  • S 1 operation When the half-pushed state is released, the focusing control is stopped.
  • the release button 115 outputs a manipulation signal for starting a photographing operation when the release button 115 is fully-pushed (S 2 operation).
  • the photographing apparatus 100 may include a manipulation member (not shown) in addition to the release button 115 .
  • the manipulation member may be, for example, up, down, left, and right keys, a power switch, or a mode dial.
  • the manipulation member transmits a manipulation signal to the DSPs 121 and 122 based on a user's manipulation.
  • the timing generator 117 outputs a timing signal to the image sensors 103 and 104 or to the CDS/AMP to control an exposure time of pixels in the image sensors 103 and 104 and reading of electric charges.
  • the DSPs 121 and 122 may be examples of a first image signal processor and a second image signal processor, and may perform as a calculating device and a controlling device according to a program.
  • the DSPs 121 and 122 control processes of components installed in the photographing apparatus 100 .
  • the DSPs 121 and 122 drive the optical systems 101 and 102 by outputting a signal to a driver (not shown) based on the focusing control or the exposure control.
  • the DSPs 121 and 122 control the components installed in the photographing apparatus 100 based on a signal transmitted from the manipulation member (not shown).
  • the DSPs 121 and 122 are respectively formed in each photographing system as shown in FIG. 1 .
  • the DSPs 121 and 122 are separately formed in order to execute signal-based commands and manipulation-based commands separately.
  • the photographing apparatus 100 of the present embodiment includes the two DSPs 121 and 122 , however, the present invention is not limited thereto.
  • the photographing apparatus of the present invention may include one DSP 221 .
  • FIG. 5 is a block diagram of a photographing apparatus 200 according to another embodiment of the present invention. That is, as shown in FIG. 5 , the one DSP 221 performs the signal-based commands and the manipulation-based commands in the photographing apparatus 200 .
  • Components of the photographing apparatus 200 other than the DSP 221 are the same as those of the photographing apparatus 100 shown in FIG. 1 , and thus, detailed descriptions of those components are not provided here.
  • the flash controller 131 generates a control signal based on a signal transmitted from the DSP 121 , and transmits the generated signal to the flash 133 .
  • the flash irradiates light onto the subject during a main photographing operation and during controlling of the focus before a main photographing operation.
  • the LCD controller 135 receives image data from, for example, an encoder LCD controller 163 , and displays the image data on the LCD 137 .
  • the LCD 137 is installed in a main body of the photographing apparatus 100 .
  • the images displayed on the LCD 137 may include an image read from the SDRAM 105 before the photographing (live view), various setting images of the photographing apparatus 100 , and captured and recorded images.
  • the LCD 137 is used as a display unit and the LCD controller 135 is used as a display driving unit, however, the present invention is not limited thereto. Instead, an organic electroluminescence (EL) display may be used as the display unit, and an organic EL controller may be used as the display driving unit.
  • EL organic electroluminescence
  • FIG. 2 is a block diagram of the DSPs 121 and 122 in the photographing apparatus 100 , according to an embodiment of the present invention.
  • Each of the DSPs 121 and 122 includes an image signal processor 141 , an auto exposure (AE)/auto focusing (AF)/auto white balance (AWB) calculator 142 , an exposure/AF operation controller 143 , the memory controller 151 , an image combining processor 152 , a correspondence/feature detector 153 , an image correlation detector 154 , a scene determiner 155 , a motion vector detector 156 , a recording medium controller 161 , an external communication controller 162 , the encoder LCD controller 163 , an image compression/extension processor 164 , a central processing unit (CPU) 171 , a serial input/output (S 10 ) 172 , and an input/output (I/O) port 173 .
  • CPU central processing unit
  • S 10 serial input/output
  • I/O input/output
  • the image signal processor 141 receives the digital signal output from the A/D converter of the image sensors 103 and 104 and performs image processing and generates an image signal.
  • the image signal processor 141 processes the image signal based on a white balance (WB) control value, a ⁇ value, and a contour emphasizing control value.
  • WB white balance
  • the image signal processor 141 calculates an AE evaluation value, an AWB evaluation value, and an AF evaluation value based on the image signal.
  • the AE/AF/AWB calculator 142 calculates a diaphragm amount or a shutter speed based on the AE evaluation value calculated by the image signal processor 141 .
  • the AE/AF/AWB calculator 142 calculates color signal gains of, for example, three primitive colors, based on the AWB evaluation value calculated by the image signal processor 141 .
  • the AE evaluation value and the AWB evaluation value are calculated by the image signal processor 141 based on brightness of the image signal.
  • the image signal processor 141 or the AE/AF/AWB calculator 142 may be an example of an exposure amount calculator.
  • the exposure/AF operation controller 143 outputs the diaphragm amount calculated by the AE/AF/AWB calculator 142 to the driver (not shown) of the optical system 101 as a control signal.
  • the driver generates a driving signal based on the control signal transmitted from the exposure/AF operation controller 143 .
  • the exposure/AF operation controller 143 controls exposure time, gain, and reading mode of the image sensors 103 and 104 .
  • the gain is used to calculate a contrast value.
  • the reading mode of the image sensors 103 and 104 denotes a signal processing mode for when image data is read from the image sensors 103 and 104 . That is, when the image of the subject is dark, addition of pixels is performed, or when the image of the subject is bright, the pixels in the image are read as they are.
  • the exposure/AF operation controller 143 may be an example of an exposure controller and an exposure amount setting unit.
  • the exposure/AF operation controller 143 when a command signal for starting the focusing control is received, the exposure/AF operation controller 143 generates a control signal for moving the focusing lens to a side and outputs the control signal to the driver (not shown).
  • the exposure/AF operation controller 143 calculates an in-focus location of the focusing lens based on the AF evaluation value calculated by the image signal processor 141 .
  • the AF evaluation value may be a contrast value of the image, and thus, it is determined that the image of the subject is in the in-focus state on the imaging surfaces of the image sensors 103 and 104 when the contrast value is a maximum (contrast detection).
  • the memory controller 151 controls reading/writing of the image data from/into the SDRAMs 105 and 106 .
  • DRAMs may be used instead of the SDRAMs 105 and 106 .
  • the image combination processor 152 may be an example of a dislocation detecting and combining unit, and may combine two images captured by the image sensors 103 and 104 at the same time or at different times. Here, a dislocation amount between the images is detected, and the image combination processor 152 combines the images based on the dislocation amount.
  • the correspondence/feature detector 153 extracts feature points from image frames, and detects correspondences between frames to compare the images with each other or to trace the correspondences.
  • the image correlation detector 154 calculates correlation between the image frames.
  • the image correlation detector 154 makes a correspondence between the images when a few feature points are detected by the correspondence/feature detector 153 or there is no correspondence between the image frames.
  • the scene detector 155 determines whether the photographing scene is about, for example, scenery, persons, or sports, from the detecting result of the correspondence/feature detector 153 or the image correlation detector 154 .
  • the recording medium controller 161 controls writing of the image data into the memory card 107 , or reading of the image data or setting information stored in the memory card 107 .
  • the external communication controller 162 transmits/receives signals to/from an external device such as a personal computer or a printer.
  • the encoder LCD controller 163 encodes image data to generate image data that may be displayed on the LCD 137 .
  • the image compression/extension processor 164 compresses the image signal into a joint photographic experts group (JPEG) format or a motion picture experts group (MPEG) format. Then, the image compression/extension processor 164 transmits the compressed image data to the recording medium controller 161 .
  • JPEG joint photographic experts group
  • MPEG motion picture experts group
  • the CPU 171 may be a calculator and controller for the DSPs 121 and 122 .
  • the CPU 171 sends/receives signals to/from outside via the SIO 172 or the I/O port 173 .
  • the photographing apparatus 100 may perform a series of processes according to hardware or software such as a computer program.
  • the photographing apparatus 100 includes two photographing systems; however, the present invention is not limited thereto.
  • the photographing apparatus of the present invention may include three or more photographing systems.
  • the photographing apparatus 100 according to the present embodiment may obtain images at the same time or at different times by using a plurality of photographing systems.
  • the photographing apparatus 100 includes a functional block having a function of calculating correlation between obtained images, a functional block having a function of recognizing objects (pattern matching) by extracting image frames including certain shapes, and a functional block having an algorithm of determining the photographing scene from the recognition result.
  • the photographing apparatus 100 of the present embodiment may perform both general image capturing sequences and stereo image capturing sequences.
  • the photographing apparatus 100 two images are captured at the same time using different exposure parameters from each other and are combined by using two photographing systems.
  • a dynamic range of the image may be enlarged by combining the images.
  • a main photographing operation is performed by applying a standard value of measured light as an exposure parameter to each of the photographing systems.
  • the image data is recorded in the recording medium in a stereo image format in a stereo image capturing mode.
  • one of the photographing systems mainly operates to output the live view image on the LCD 137
  • the other photographing system mainly operates to measure light for the main photographing operation.
  • FIGS. 3A and 3B are flowcharts illustrating operations of the photographing apparatus 100 in the dynamic range enlarging mode, according to an embodiment of the present invention.
  • L denotes a left photographing system of the photographing apparatus 100 (hereinafter, referred to as an L system)
  • R denotes a right photographing system of the photographing apparatus 100 (hereinafter, referred to as an R system). Both of the two photographing systems may have the same performances, such as the definition as each other.
  • the L system drives an electronic view finder (EVF) in an image outputting mode, and starts displaying the live view on the LCD 137 (operation S 101 ).
  • the R system drives the light measuring mode, and starts the light measuring operation by using the image sensor of the R system (operation S 101 ).
  • a frame rate of the R system is set to be equal to that of the L system or higher.
  • the exposure amount of the EVF may be set according to a reference light value measured by the R system.
  • the L system extracts feature points from, for example, an nth frame image (operation S 102 ). Meanwhile, the R system calculates a histogram from a standard light measuring result of the n-th frame image (operation S 102 ).
  • the L system extracts feature points from an (n+1)th frame image, and traces the feature points by comparing the feature points extracted from the (n+1)th frame with the feature points extracted in the operation S 102 (operation S 103 ). Meanwhile, the R system calculates a histogram from a light measuring result based on a highlight reference in the (n+1)th frame image (operation S 103 ).
  • the L system determines the photographing scene from the extracted feature points or the tracing result of the feature points (operation S 104 ). For example, the L system determines whether the photographing scene coincides with the photographing scenes such as scenery, persons, or sports stored in the photographing apparatus 100 in advance, and outputs the determination of the photographing scene when there is a corresponding scene in the photographing apparatus 100 . Thus, a photographing mode corresponding to the scene may be determined for the main photographing operation, and photographing operations suitable for the scene may be performed accordingly.
  • the L system may perform a face recognition process in an (n+2)th frame image by using the extracted feature points or the tracing result of the feature points.
  • the R system calculates a histogram from a light measuring result based on dark reference (based on shadow reference) from the (n+2)th frame image (operation S 104 ).
  • the dynamic range of the main photographing operation is calculated based on the histograms of the standard light measuring, the highlight-based light measuring, and the dark-based light measuring, which are calculated by the R system, and then the exposure parameters for the L and R systems in the main photographing operation may be calculated.
  • the photographing mode of the main photographing operation corresponding to the determined scene is determined (operation S 105 ).
  • the L system sets the exposure parameter calculated based on the highlight reference
  • the R system sets the exposure parameter calculated based on the dark reference (operation S 106 ). Therefore, the image captured in the main photographing operation may have a wider dynamic range than that of an image captured using only one imaging device.
  • the image data captured based on the highlight reference by the L system and the image data captured based on the dark reference by the R system are simultaneously input into the image signal processor 141 (operation S 108 ).
  • the image correlation detector 154 performs a pattern matching operation of the image data obtained through the L and R systems (operation S 109 ). In more detail, the image correlation detector 154 calculates a correlation between the two images to detect a dislocation between the two images.
  • the image captured based on the highlight reference by the L system and the image captured based on the dark reference by the R system are combined with each other (operation S 110 ). At this time, the two images are combined, removing the dislocation therebetween, and thus, a clear image may be obtained. If necessary, portions of the combined image not required are cut off (operation S 111 ).
  • the images obtained by the L and R systems are slightly different from each other even when the L and R systems are close to each other. Therefore, when the images are combined, removing the detected dislocation between the images by correlation calculation, a general flat image may be obtained even when the images obtained by the L and R systems are slightly different from each other.
  • images captured at different exposure parameters are obtained simultaneously, and an image having a wide dynamic range may be obtained by the combination of the captured images.
  • one of the two photographing systems mainly operates to output the image to the LCD 137 , and the other photographing system mainly performs the light measuring for the main photographing operation. Therefore, unlike the case where the live view image is obtained and performing the light measuring is performed using only one photographing system, an image exclusively for the live view may be obtained separately. Thus, repetition of brightness and darkness in the image and discontinuity in displaying the live view of the subject that may be caused by the light measuring operation may be prevented.
  • the L and R systems have the same definition as each other, however, they may have different definitions from each other provided that the qualities of the images obtained from the L and R systems become the same through image processing.
  • the R system performs the light measuring in an order of the standard light measuring, the light measuring based on highlight reference, and the light measuring based on the dark reference, however, the present invention is not limited thereto.
  • FIGS. 4A and 4B are flowcharts illustrating operations of the photographing apparatus 100 in the stereo image capturing mode.
  • the photographing apparatus 100 After displaying the live view of the subject on the LCD 137 , the photographing apparatus 100 performs a main photographing operation when the release button 115 is completely pushed.
  • L denotes the left photographing system of the photographing apparatus 100 (hereinafter, referred to as an L system)
  • R denotes the right photographing system of the photographing apparatus 100 (hereinafter, referred to as an R system).
  • L system the left photographing system of the photographing apparatus 100
  • R system the right photographing system of the photographing apparatus 100
  • the L system drives an EVF in an image outputting mode, and starts displaying the live view on the LCD 137 (operation S 201 ).
  • the R system is driven in the light measuring mode, and starts the light measuring operation by using the image sensor of the R system (operation S 201 ).
  • a frame rate of the R system is set to be equal to that of the L system or higher.
  • the exposure amount of the EVF may be set according to a reference light value measured by the R system.
  • the L system obtains an nth frame image, and displays the image as the live view on the LCD 137 (operation S 202 ). Meanwhile, the R system calculates a histogram from a standard measuring result of the nth frame image (operation S 202 ).
  • the L system obtains an (n+1)th frame image, and displays the obtained image as the live view on the LCD 137 (operation S 203 ).
  • the R system calculates a histogram from a light measuring result based on highlight reference in the (n+1)th frame image (operation S 203 ).
  • the L system obtains an (n+2)th frame image, and displays the obtained image as the live view on the LCD 137 (operation S 204 ). Meanwhile, the R system calculates a histogram from the light measuring result based on dark reference (based on shadow reference) from the (n+2)th frame image (operation S 204 ).
  • the dynamic range of the main photographing operation is calculated based on the histograms of the standard light measuring, the highlight-based light measuring, and the dark-based light measuring, which are calculated by the R system, and then the exposure parameters for the L and R systems in the main photographing operation may be calculated (operation S 205 ).
  • the exposure parameters of the L and R systems are set by using the standard light measuring value (operation S 206 ).
  • the exposure parameter may be determined without using the standard light measuring value if the L and R systems have the same exposure value and the exposure parameter is suitable for both of the L and R systems.
  • the exposure parameter for the main photographing operation may be set by the light measured by the R system. Otherwise, operations S 202 through S 206 may be omitted, and the exposure parameter may be set by using the standard light measuring value.
  • the image data captured by the L system and the image data captured by the R system are simultaneously input into the image signal processor 141 (operation S 208 ).
  • the image correlation detector 154 performs a pattern matching operation of the image data obtained through the L and R systems (operation S 209 ). In more detail, the image correlation detector 154 calculates a correlation between the two images to detect a dislocation between the two images. In addition, a parameter for generating a stereo image may be calculated based on the two images captured by the R and L systems, the dislocation amount between the two images, and characteristics of the photographing apparatus 100 (operation S 210 ).
  • the image correlation detector 154 detects the dislocation through the correlation calculation in order to calculate the parameter for generating the stereo image, and accordingly, the image data may be recorded in the memory card 107 in the stereo image format by which the stereo image may be reproduced.
  • one of the two photographing systems mainly operates to output the image to the LCD 137 , and the other photographing system mainly performs the light measuring for the main photographing operation. Therefore, unlike the case where the live view image is obtained and the light measuring is performed using only one photographing system, an image exclusively for the live view may be obtained separately. Thus, repetition of brightness and darkness in the image and discontinuity in displaying the live view of the subject that may be caused by the light measuring operation may be prevented.
  • the main photographing is performed by using the standard light measuring value only; however, after starting the main photographing operation, the image data of a certain frame may be captured by the L and R systems simultaneously based on the highlight reference, and then, the image data of the next frame may be captured by the L and R systems simultaneously based on the dark reference. Then, in each of the L and R systems the image data based on the highlight reference and the image data based on the dark reference may be combined to generate an image having wide dynamic range.
  • the moving picture may be stably displayed on an LCD before performing a main photographing operation of a still image, and an exposure amount required to perform the main photographing operation of the still image may be ensured.
  • the device described herein may comprise a processor, a memory for storing program data and executing it, a permanent storage such as a disk drive, a communications port for handling communications with external devices, and user interface devices, including a display, keys, etc.
  • software modules When software modules are involved, these software modules may be stored as program instructions or computer readable codes executable on the processor on a computer-readable media such as read-only memory (ROM), random-access memory (RAM), CD-ROMs, magnetic tapes, floppy disks, and optical data storage devices.
  • ROM read-only memory
  • RAM random-access memory
  • CD-ROMs compact discs
  • magnetic tapes magnetic tapes
  • floppy disks floppy disks
  • optical data storage devices optical data storage devices.
  • the computer readable recording medium can also be distributed over network coupled computer systems so that the computer readable code is stored and executed in a distributed fashion. This media can be read by the computer, stored in the memory, and executed by the processor.
  • the present invention may be described in terms of functional block components and various processing steps. Such functional blocks may be realized by any number of hardware and/or software components configured to perform the specified functions.
  • the present invention may employ various integrated circuit components, e.g., memory elements, processing elements, logic elements, look-up tables, and the like, which may carry out a variety of functions under the control of one or more microprocessors or other control devices.
  • the elements of the present invention are implemented using software programming or software elements the invention may be implemented with any programming or scripting language such as C, C++, Java, assembler, or the like, with the various algorithms being implemented with any combination of data structures, objects, processes, routines or other programming elements.
  • Functional aspects may be implemented in algorithms that execute on one or more processors.
  • the present invention could employ any number of conventional techniques for electronics configuration, signal processing and/or control, data processing and the like.
  • the words “mechanism” and “element” are used broadly and are not limited to mechanical or physical embodiments, but can include software routines in conjunction with processors, etc.

Abstract

A photographing apparatus and method for dynamic range adjustment and stereography. The photographing apparatus includes a first imaging device for converting a light of a subject received through a first optical system into an electric signal; a second imaging device for converting a light of the subject received through a second optical system into an electric signal; a first image signal processor for generating an image signal for live view based on the electric signal output from the first imaging device before a photographing operation of a still image; an exposure controller for controlling an exposure so as to perform a step exposure in the second imaging device before the photographing of the still image; and an exposure calculator for calculating an exposure amount in the photographing operation of the still image based on the electric signal converted in the second imaging device obtained through the step exposure.

Description

CROSS-REFERENCE TO RELATED PATENT APPLICATION
This application claims the benefit of Japanese Patent Application No. 2008-303837, filed on Nov. 28, 2008, in the Japanese Intellectual Property Office, the disclosure of which is incorporated herein in its entirety by reference.
BACKGROUND
The present invention relates to a photographing apparatus and method for dynamic range adjustment and stereography.
In scenes captured by a photographing apparatus such as a digital still camera or a video camera, there may be a wide brightness ratio between the brightest portion and the darkest portion of a subject. This ratio of brightest to darkest is also called the dynamic range of the image.
When photographing of the subject is performed using a photographing apparatus having a narrow allowable dynamic range, the subject may not be photographed well. At this time, since the photographing apparatus is not in an appropriate exposure state, the captured image may include dark or bright smeared portions.
When the subject is photographed by a photographing apparatus having a narrow allowable dynamic range, the dynamic range may be enlarged using the following operations. For example, the dynamic range may be enlarged by (1) non-linear treatment such as gamma or knee operation, (2) using a non-linear photographing apparatus, (3) combining images obtained from imaging devices having properties different from each other, or (4) image signal processing or combining images obtained from one imaging device.
However, since the dynamic range of original image data input into the photographing apparatus is narrow, there is a limitation in enlarging the dynamic range using the operations (1) through (4) mentioned above.
In addition, according to a conventional method of enlarging a dynamic range by combining a plurality of images, exposures of which are different from each other, images captured at different times from each other may be combined. Thus, it is difficult to combine images obtained by photographing a moving subject or captured during a hand-shake to produce an image having high image quality. In addition, in the above case, each of the images may be captured after changing exposure, and thus, brightness of an electronic view finder (EVF) or a liquid crystal display (LCD) in the live view display may be repeatedly changed between dark and bright states, or continuity of displaying the subject in the live view display may not be ensured.
SUMMARY
The present invention provides a photographing apparatus and method, in which an exposure amount required to successfully perform a main photographing operation for a still image is ensured while displaying a moving picture image stably before the main photographing operation for the still image.
According to an aspect of the present invention, there is provided a photographing apparatus including: a first imaging device for converting light from a subject received through a first optical system into an electric signal; a second imaging device for converting light from the subject received through a second optical system into an electric signal; a first image signal processor for generating an image signal for a live view based on the electric signal output from the first imaging device before a photographing operation of a still image; an exposure controller for controlling an exposure so as to perform a step exposure in the second imaging device before the photographing of the still image; and an exposure calculator for calculating an exposure amount during the photographing operation of the still image based on the electric signal converted in the second imaging device obtained through the step exposure.
The photographing apparatus may further include: an exposure amount setting unit for setting an exposure amount of the first or second imaging device based on highlight reference and an exposure amount of the other imaging devices based on dark reference at the same time based on the calculated exposure amount; and a second image processor for generating a first image signal for a still image based on the electric signal converted by the first imaging device and a second image signal for a still image based on the electric signal converted by the second imaging device at the same time when the photographing operation of the still image is performed based on the set exposure amount.
The photographing apparatus may further include: a dislocation detector for detecting a dislocation between the first and second image signals; and a combining unit for combining the first and second image signals with each other using the detected dislocation.
The photographing apparatus may further include: an exposure amount setting unit for setting exposure amounts of the first and second imaging devices to be the same as each other at the same timing based on the calculated exposure amount; and a second image processor for generating a first image signal for a still image based on the electric signal converted by the first imaging device and a second image signal for a still image based on the electric signal converted by the second imaging device at the same time when the photographing operation of the still image is performed based on the set exposure amount.
The photographing apparatus may further include: a dislocation detector for detecting a dislocation between the first and second image signals; and a combining unit for combining the first and second image signals with each other using the detected dislocation.
According to another aspect of the present invention, there is provided a photographing method including: converting light from a subject received through a first optical system into an electric signal by a first imaging device; converting light from the subject received through a second optical system into an electric signal by a second imaging device; generating an image signal for a live view based on the electric signal converted by the first imaging device before a photographing operation of a still image; controlling an exposure to be a step exposure in the second imaging device before the photographing operation of the still image; and calculating an exposure amount required in the photographing operation of the still image based on the electric signal converted by the second imaging device, which is obtained by the step exposure.
BRIEF DESCRIPTION OF THE DRAWINGS
The above and other features and advantages of the present invention will become more apparent by describing in detail exemplary embodiments thereof with reference to the attached drawings in which:
FIG. 1 is a block diagram of a photographing apparatus according to an embodiment of the present invention;
FIG. 2 is a block diagram of a digital signal processor (DSP) in the photographing apparatus of FIG. 1, according to an embodiment of the present invention;
FIGS. 3A and 3B are flowcharts illustrating operations of the photographing apparatus of FIG. 1 in a dynamic range enlarging mode, according to an embodiment of the present invention;
FIGS. 4A and 4B are flowcharts illustrating operations of the photographing apparatus of FIG. 1 in a stereo image photographing mode, according to an embodiment of the present invention; and
FIG. 5 is a block diagram of a photographing apparatus according to another embodiment of the present invention.
DETAILED DESCRIPTION OF THE EMBODIMENTS
Hereinafter, the present invention will be described in detail by explaining embodiments of the invention with reference to the attached drawings. Like reference numerals in the drawings denote like elements.
A photographing apparatus 100 according to an embodiment of the present invention will be described as follows.
FIG. 1 is a block diagram of the photographing apparatus 100 according to the embodiment of the present invention.
The photographing apparatus 100 includes two photographing systems each of which including an optical system and an imaging system, and each of the photographing systems may separately photograph a subject to obtain an image. For example, when a live view image is displayed, an image displayed as the live view image may be obtained through one photographing system, and the other photographing system may measure light. Otherwise, the photographing systems may be arranged parallel to each other to photograph the subject at the same time, and thus, a stereo image (a three-dimensional image) may be generated.
The photographing apparatus 100 includes optical systems 101 and 102, image sensors 103 and 104, synchronous random access memories (SDRAMs) 105 and 106, a memory card 107, a read only memory 109, a release button 115, a timing generator 117, digital signal processors 121 and 122, a flash controller 131, a flash 133, a liquid crystal display (LCD) controller 135, and an LCD 137.
The optical systems 101 and 102 are respectively examples of a first optical system and a second optical system. The optical systems 101 and 102 form image of external light on the image sensors 103 and 104, respectively transmitting light from the subject toward the image sensors 103 and 104. The light from the optical system 101 reaches the image sensor 103, and the light from the optical system 102 reaches the image sensor 104.
Each of the optical systems 101 and 102 may include a zoom lens, a diaphragm, and a focusing lens. The zoom lens changes a viewing angle by changing a focal length, and the diaphragm adjusts an amount of the light transmitting therethrough. The focusing lens focuses an image of the subject on an imaging surface of the image sensor 103 or 104 by moving from a side to another side, or vice versa.
The image sensors 103 and 104 are examples of a first imaging device and a second imaging device (photoelectric conversion device), and may include photoelectric conversion devices that convert optical information incident from the optical systems 101 and 102 into electric signals. The image sensors 103 and 104 may generate the electric signals according to the received light.
Charge coupled device (CCD) sensors or complementary metal oxide semiconductor (CMOS) sensors may be used as the image sensors 103 and 104.
In addition, a mechanical shutter (not shown) that may block light when photographing is not performed and may transmit light when photographing is performed may be used to control exposure times of the image sensors 103 and 104. However, the present invention is not limited thereto, that is, an electronic shutter (not shown) may be used. In addition, the mechanical shutter or the electronic shutter may be operated by a switch connected to the DSPs 121 and 122, such as the release button 115 (a manipulation member).
Each of the image sensors 103 and 104 respectively further includes a correlated double sampler (CDS)/amplifier (AMP), and an analog/digital (A/D) converter.
The CDS/AMP removes low-frequency noise included in the electric signal output from the image sensors 103 and 104, and at the same time, each may amplify a respective electric signal to a predetermined level.
The A/D converter converts the electric signal output from the CDS/AMP into a digital signal. The A/D converter outputs the digital signal to an image signal processor 141.
The SDRAMs 105 and 106 temporarily store image data of captured images. The SDRAMs 105 and 106 may store image data of a plurality of images. In addition, the SDRAMs 105 and 106 are respectively connected to the DSPs 121 and 122. Reading/writing of image data from/to the SDRAMs 105 and 106 are controlled by a memory controller 151.
In addition, the SDRAM 105 may include a region as video RAM (VRAM). The VRAM may be used for displaying images, and may include a plurality of channels. The VRAMs may simultaneously perform operations of inputting image data therein and outputting image data to the LCD controller 135. A definition or a maximum color representation number may depend on a capacity of the VRAM.
The memory card 107 may include a semiconductor memory medium such as a flash memory. Image data generated by a photographing operation is recorded in the memory card 107 or is read from the memory card 107. In addition, the memory card 107 is not limited to the flash memory, but may include an optical disc (a compact disc (CD), a digital versatile disc (DVD), or a blue-ray disc), a photomagnetic disc, or a magnetic disc. The memory card 107 may be detachable from the photographing apparatus 100.
The ROM 109 stores an operating program for the DSPs 121 and 122.
The release button 115 may operate in a half-pushed state, a fully-pushed state, and a released state. The release button 115 outputs a manipulation signal for starting focusing control when the release button 115 is half-pushed (S1 operation). When the half-pushed state is released, the focusing control is stopped. In addition, the release button 115 outputs a manipulation signal for starting a photographing operation when the release button 115 is fully-pushed (S2 operation).
In addition, the photographing apparatus 100 may include a manipulation member (not shown) in addition to the release button 115. The manipulation member may be, for example, up, down, left, and right keys, a power switch, or a mode dial. The manipulation member transmits a manipulation signal to the DSPs 121 and 122 based on a user's manipulation.
The timing generator 117 outputs a timing signal to the image sensors 103 and 104 or to the CDS/AMP to control an exposure time of pixels in the image sensors 103 and 104 and reading of electric charges.
The DSPs 121 and 122 may be examples of a first image signal processor and a second image signal processor, and may perform as a calculating device and a controlling device according to a program. The DSPs 121 and 122 control processes of components installed in the photographing apparatus 100.
The DSPs 121 and 122 drive the optical systems 101 and 102 by outputting a signal to a driver (not shown) based on the focusing control or the exposure control. In addition, the DSPs 121 and 122 control the components installed in the photographing apparatus 100 based on a signal transmitted from the manipulation member (not shown). In addition, according to the present embodiment, the DSPs 121 and 122 are respectively formed in each photographing system as shown in FIG. 1. The DSPs 121 and 122 are separately formed in order to execute signal-based commands and manipulation-based commands separately.
The photographing apparatus 100 of the present embodiment includes the two DSPs 121 and 122, however, the present invention is not limited thereto. For example, as shown in FIG. 5, the photographing apparatus of the present invention may include one DSP 221. FIG. 5 is a block diagram of a photographing apparatus 200 according to another embodiment of the present invention. That is, as shown in FIG. 5, the one DSP 221 performs the signal-based commands and the manipulation-based commands in the photographing apparatus 200. Components of the photographing apparatus 200 other than the DSP 221 are the same as those of the photographing apparatus 100 shown in FIG. 1, and thus, detailed descriptions of those components are not provided here.
The flash controller 131 generates a control signal based on a signal transmitted from the DSP 121, and transmits the generated signal to the flash 133. The flash irradiates light onto the subject during a main photographing operation and during controlling of the focus before a main photographing operation.
The LCD controller 135 receives image data from, for example, an encoder LCD controller 163, and displays the image data on the LCD 137. The LCD 137 is installed in a main body of the photographing apparatus 100.
The images displayed on the LCD 137 may include an image read from the SDRAM 105 before the photographing (live view), various setting images of the photographing apparatus 100, and captured and recorded images. In the present embodiment, the LCD 137 is used as a display unit and the LCD controller 135 is used as a display driving unit, however, the present invention is not limited thereto. Instead, an organic electroluminescence (EL) display may be used as the display unit, and an organic EL controller may be used as the display driving unit.
Next, the DSPs 121 and 122 of the photographing apparatus 100 according to the present embodiment will be described as follows. FIG. 2 is a block diagram of the DSPs 121 and 122 in the photographing apparatus 100, according to an embodiment of the present invention.
Each of the DSPs 121 and 122 includes an image signal processor 141, an auto exposure (AE)/auto focusing (AF)/auto white balance (AWB) calculator 142, an exposure/AF operation controller 143, the memory controller 151, an image combining processor 152, a correspondence/feature detector 153, an image correlation detector 154, a scene determiner 155, a motion vector detector 156, a recording medium controller 161, an external communication controller 162, the encoder LCD controller 163, an image compression/extension processor 164, a central processing unit (CPU) 171, a serial input/output (S10) 172, and an input/output (I/O) port 173.
The image signal processor 141 receives the digital signal output from the A/D converter of the image sensors 103 and 104 and performs image processing and generates an image signal. The image signal processor 141 processes the image signal based on a white balance (WB) control value, aγ value, and a contour emphasizing control value. In addition, the image signal processor 141 calculates an AE evaluation value, an AWB evaluation value, and an AF evaluation value based on the image signal.
The AE/AF/AWB calculator 142 calculates a diaphragm amount or a shutter speed based on the AE evaluation value calculated by the image signal processor 141. The AE/AF/AWB calculator 142 calculates color signal gains of, for example, three primitive colors, based on the AWB evaluation value calculated by the image signal processor 141. The AE evaluation value and the AWB evaluation value are calculated by the image signal processor 141 based on brightness of the image signal. The image signal processor 141 or the AE/AF/AWB calculator 142 may be an example of an exposure amount calculator.
The exposure/AF operation controller 143 outputs the diaphragm amount calculated by the AE/AF/AWB calculator 142 to the driver (not shown) of the optical system 101 as a control signal. The driver generates a driving signal based on the control signal transmitted from the exposure/AF operation controller 143. The exposure/AF operation controller 143 controls exposure time, gain, and reading mode of the image sensors 103 and 104. The gain is used to calculate a contrast value. Here, the reading mode of the image sensors 103 and 104 denotes a signal processing mode for when image data is read from the image sensors 103 and 104. That is, when the image of the subject is dark, addition of pixels is performed, or when the image of the subject is bright, the pixels in the image are read as they are. The exposure/AF operation controller 143 may be an example of an exposure controller and an exposure amount setting unit.
In addition, when a command signal for starting the focusing control is received, the exposure/AF operation controller 143 generates a control signal for moving the focusing lens to a side and outputs the control signal to the driver (not shown). The exposure/AF operation controller 143 calculates an in-focus location of the focusing lens based on the AF evaluation value calculated by the image signal processor 141. The AF evaluation value may be a contrast value of the image, and thus, it is determined that the image of the subject is in the in-focus state on the imaging surfaces of the image sensors 103 and 104 when the contrast value is a maximum (contrast detection).
The memory controller 151 controls reading/writing of the image data from/into the SDRAMs 105 and 106. Here, DRAMs may be used instead of the SDRAMs 105 and 106.
The image combination processor 152 may be an example of a dislocation detecting and combining unit, and may combine two images captured by the image sensors 103 and 104 at the same time or at different times. Here, a dislocation amount between the images is detected, and the image combination processor 152 combines the images based on the dislocation amount.
The correspondence/feature detector 153 extracts feature points from image frames, and detects correspondences between frames to compare the images with each other or to trace the correspondences.
The image correlation detector 154 calculates correlation between the image frames. The image correlation detector 154 makes a correspondence between the images when a few feature points are detected by the correspondence/feature detector 153 or there is no correspondence between the image frames.
The scene detector 155 determines whether the photographing scene is about, for example, scenery, persons, or sports, from the detecting result of the correspondence/feature detector 153 or the image correlation detector 154.
The recording medium controller 161 controls writing of the image data into the memory card 107, or reading of the image data or setting information stored in the memory card 107.
The external communication controller 162 transmits/receives signals to/from an external device such as a personal computer or a printer.
The encoder LCD controller 163 encodes image data to generate image data that may be displayed on the LCD 137.
The image compression/extension processor 164 compresses the image signal into a joint photographic experts group (JPEG) format or a motion picture experts group (MPEG) format. Then, the image compression/extension processor 164 transmits the compressed image data to the recording medium controller 161.
The CPU 171 may be a calculator and controller for the DSPs 121 and 122. The CPU 171 sends/receives signals to/from outside via the SIO 172 or the I/O port 173.
In addition, the photographing apparatus 100 may perform a series of processes according to hardware or software such as a computer program.
As described above, the photographing apparatus 100 according to the present embodiment includes two photographing systems; however, the present invention is not limited thereto. For example, the photographing apparatus of the present invention may include three or more photographing systems. The photographing apparatus 100 according to the present embodiment may obtain images at the same time or at different times by using a plurality of photographing systems.
In addition, the photographing apparatus 100 according to the present embodiment includes a functional block having a function of calculating correlation between obtained images, a functional block having a function of recognizing objects (pattern matching) by extracting image frames including certain shapes, and a functional block having an algorithm of determining the photographing scene from the recognition result. In addition, the photographing apparatus 100 of the present embodiment may perform both general image capturing sequences and stereo image capturing sequences.
Operations of the photographing apparatus 100 according to the present embodiment will be described as follows.
In the photographing apparatus 100, two images are captured at the same time using different exposure parameters from each other and are combined by using two photographing systems. In addition, in a mode of outputting a two-dimensional image, a dynamic range of the image may be enlarged by combining the images.
Moreover, in a mode of outputting a stereo (three-dimensional) image, a main photographing operation is performed by applying a standard value of measured light as an exposure parameter to each of the photographing systems. The image data is recorded in the recording medium in a stereo image format in a stereo image capturing mode.
In addition, when the live view is displayed before a main photographing operation, one of the photographing systems mainly operates to output the live view image on the LCD 137, and the other photographing system mainly operates to measure light for the main photographing operation.
[Dynamic Range Enlarging Mode]
Next, a case where the photographing apparatus 100 is used in a dynamic range enlarging mode will be described. FIGS. 3A and 3B are flowcharts illustrating operations of the photographing apparatus 100 in the dynamic range enlarging mode, according to an embodiment of the present invention.
Here, after displaying the live view of the subject on the LCD 137, a main photographing operation is performed when the release button 115 is fully-pushed. In addition, in FIGS. 3A and 3B, L denotes a left photographing system of the photographing apparatus 100 (hereinafter, referred to as an L system), and R denotes a right photographing system of the photographing apparatus 100 (hereinafter, referred to as an R system). Both of the two photographing systems may have the same performances, such as the definition as each other.
First, the L system drives an electronic view finder (EVF) in an image outputting mode, and starts displaying the live view on the LCD 137 (operation S101). The R system drives the light measuring mode, and starts the light measuring operation by using the image sensor of the R system (operation S101). Here, a frame rate of the R system is set to be equal to that of the L system or higher. Thus, the exposure amount of the EVF may be set according to a reference light value measured by the R system.
Next, the L system extracts feature points from, for example, an nth frame image (operation S102). Meanwhile, the R system calculates a histogram from a standard light measuring result of the n-th frame image (operation S102).
In addition, the L system extracts feature points from an (n+1)th frame image, and traces the feature points by comparing the feature points extracted from the (n+1)th frame with the feature points extracted in the operation S102 (operation S103). Meanwhile, the R system calculates a histogram from a light measuring result based on a highlight reference in the (n+1)th frame image (operation S103).
Then, the L system determines the photographing scene from the extracted feature points or the tracing result of the feature points (operation S104). For example, the L system determines whether the photographing scene coincides with the photographing scenes such as scenery, persons, or sports stored in the photographing apparatus 100 in advance, and outputs the determination of the photographing scene when there is a corresponding scene in the photographing apparatus 100. Thus, a photographing mode corresponding to the scene may be determined for the main photographing operation, and photographing operations suitable for the scene may be performed accordingly. The L system may perform a face recognition process in an (n+2)th frame image by using the extracted feature points or the tracing result of the feature points.
On the other hand, the R system calculates a histogram from a light measuring result based on dark reference (based on shadow reference) from the (n+2)th frame image (operation S104).
Next, the dynamic range of the main photographing operation is calculated based on the histograms of the standard light measuring, the highlight-based light measuring, and the dark-based light measuring, which are calculated by the R system, and then the exposure parameters for the L and R systems in the main photographing operation may be calculated. In addition, the photographing mode of the main photographing operation corresponding to the determined scene is determined (operation S105).
In addition, the L system sets the exposure parameter calculated based on the highlight reference, and the R system sets the exposure parameter calculated based on the dark reference (operation S106). Therefore, the image captured in the main photographing operation may have a wider dynamic range than that of an image captured using only one imaging device.
In addition, it is determined whether the release button 115 is in a pressed state when the exposure parameters are set (operation S107). When the release button 115 is determined to be in the pressed state, the process goes to operation S108. Otherwise, the operations S102 through S106 are repeated to reset the light measuring and the exposure parameters until the release button 115 is determined to be in the pressed stated.
When the release button 115 is determined to be in a pressed state, the image data captured based on the highlight reference by the L system and the image data captured based on the dark reference by the R system are simultaneously input into the image signal processor 141 (operation S108).
Next, the image correlation detector 154 performs a pattern matching operation of the image data obtained through the L and R systems (operation S109). In more detail, the image correlation detector 154 calculates a correlation between the two images to detect a dislocation between the two images.
In addition, the image captured based on the highlight reference by the L system and the image captured based on the dark reference by the R system are combined with each other (operation S110). At this time, the two images are combined, removing the dislocation therebetween, and thus, a clear image may be obtained. If necessary, portions of the combined image not required are cut off (operation S111).
Since the L and R systems are located at different angles with respect to the same subject, the images obtained by the L and R systems are slightly different from each other even when the L and R systems are close to each other. Therefore, when the images are combined, removing the detected dislocation between the images by correlation calculation, a general flat image may be obtained even when the images obtained by the L and R systems are slightly different from each other. In addition, according to the present embodiment, images captured at different exposure parameters are obtained simultaneously, and an image having a wide dynamic range may be obtained by the combination of the captured images.
In addition, one of the two photographing systems mainly operates to output the image to the LCD 137, and the other photographing system mainly performs the light measuring for the main photographing operation. Therefore, unlike the case where the live view image is obtained and performing the light measuring is performed using only one photographing system, an image exclusively for the live view may be obtained separately. Thus, repetition of brightness and darkness in the image and discontinuity in displaying the live view of the subject that may be caused by the light measuring operation may be prevented.
In addition, in the above description, the L and R systems have the same definition as each other, however, they may have different definitions from each other provided that the qualities of the images obtained from the L and R systems become the same through image processing. In addition, the R system performs the light measuring in an order of the standard light measuring, the light measuring based on highlight reference, and the light measuring based on the dark reference, however, the present invention is not limited thereto.
[Stereo Image Capturing Mode]
Next, the photographing apparatus 100 operating in the stereo image capturing mode will be described as follows.
FIGS. 4A and 4B are flowcharts illustrating operations of the photographing apparatus 100 in the stereo image capturing mode.
Here, after displaying the live view of the subject on the LCD 137, the photographing apparatus 100 performs a main photographing operation when the release button 115 is completely pushed.
In addition, as described with reference to FIGS. 3A and 3B, L denotes the left photographing system of the photographing apparatus 100 (hereinafter, referred to as an L system), and R denotes the right photographing system of the photographing apparatus 100 (hereinafter, referred to as an R system). Both of the two optical imaging systems may have the same performances, such as the same definition as each other.
First, the L system drives an EVF in an image outputting mode, and starts displaying the live view on the LCD 137 (operation S201).
The R system is driven in the light measuring mode, and starts the light measuring operation by using the image sensor of the R system (operation S201). Here, a frame rate of the R system is set to be equal to that of the L system or higher. Thus, the exposure amount of the EVF may be set according to a reference light value measured by the R system.
Next, the L system obtains an nth frame image, and displays the image as the live view on the LCD 137 (operation S202). Meanwhile, the R system calculates a histogram from a standard measuring result of the nth frame image (operation S202).
In addition, the L system obtains an (n+1)th frame image, and displays the obtained image as the live view on the LCD 137 (operation S203). Meanwhile, the R system calculates a histogram from a light measuring result based on highlight reference in the (n+1)th frame image (operation S203).
Next, the L system obtains an (n+2)th frame image, and displays the obtained image as the live view on the LCD 137 (operation S204). Meanwhile, the R system calculates a histogram from the light measuring result based on dark reference (based on shadow reference) from the (n+2)th frame image (operation S204).
Next, the dynamic range of the main photographing operation is calculated based on the histograms of the standard light measuring, the highlight-based light measuring, and the dark-based light measuring, which are calculated by the R system, and then the exposure parameters for the L and R systems in the main photographing operation may be calculated (operation S205).
In addition, the exposure parameters of the L and R systems are set by using the standard light measuring value (operation S206). However, the exposure parameter may be determined without using the standard light measuring value if the L and R systems have the same exposure value and the exposure parameter is suitable for both of the L and R systems. The exposure parameter for the main photographing operation may be set by the light measured by the R system. Otherwise, operations S202 through S206 may be omitted, and the exposure parameter may be set by using the standard light measuring value.
Next, it is determined whether the release button 115 is determined to be in a pressed state when the exposure parameter is set (operation S207). When the release button 115 is determined to be in the pressed state, the process goes to operation S208. Otherwise, the operations S202 through S206 are repeated to reset the light measuring and the exposure parameters until the release button 115 is determined to be in the pressed state.
When the release button 115 is determined to be in the pressed state, the image data captured by the L system and the image data captured by the R system are simultaneously input into the image signal processor 141 (operation S208).
Next, the image correlation detector 154 performs a pattern matching operation of the image data obtained through the L and R systems (operation S209). In more detail, the image correlation detector 154 calculates a correlation between the two images to detect a dislocation between the two images. In addition, a parameter for generating a stereo image may be calculated based on the two images captured by the R and L systems, the dislocation amount between the two images, and characteristics of the photographing apparatus 100 (operation S210).
Next, the image data and the parameter are recorded in the memory card 107 in the stereo image format (operation S211).
Since the L and R systems are located at different angles with respect to the same subject, images captured by the L and R systems are slightly different from each other. Therefore, the image correlation detector 154 detects the dislocation through the correlation calculation in order to calculate the parameter for generating the stereo image, and accordingly, the image data may be recorded in the memory card 107 in the stereo image format by which the stereo image may be reproduced.
In addition, one of the two photographing systems mainly operates to output the image to the LCD 137, and the other photographing system mainly performs the light measuring for the main photographing operation. Therefore, unlike the case where the live view image is obtained and the light measuring is performed using only one photographing system, an image exclusively for the live view may be obtained separately. Thus, repetition of brightness and darkness in the image and discontinuity in displaying the live view of the subject that may be caused by the light measuring operation may be prevented.
In addition, in the stereo image capturing mode, the main photographing is performed by using the standard light measuring value only; however, after starting the main photographing operation, the image data of a certain frame may be captured by the L and R systems simultaneously based on the highlight reference, and then, the image data of the next frame may be captured by the L and R systems simultaneously based on the dark reference. Then, in each of the L and R systems the image data based on the highlight reference and the image data based on the dark reference may be combined to generate an image having wide dynamic range.
According to the present invention, the moving picture may be stably displayed on an LCD before performing a main photographing operation of a still image, and an exposure amount required to perform the main photographing operation of the still image may be ensured.
While the present invention has been particularly shown and described with reference to exemplary embodiments thereof, it will be understood by those of ordinary skill in the art that various changes in form and details may be made therein without departing from the spirit and scope of the present invention as defined by the following claims.
The device described herein may comprise a processor, a memory for storing program data and executing it, a permanent storage such as a disk drive, a communications port for handling communications with external devices, and user interface devices, including a display, keys, etc. When software modules are involved, these software modules may be stored as program instructions or computer readable codes executable on the processor on a computer-readable media such as read-only memory (ROM), random-access memory (RAM), CD-ROMs, magnetic tapes, floppy disks, and optical data storage devices. The computer readable recording medium can also be distributed over network coupled computer systems so that the computer readable code is stored and executed in a distributed fashion. This media can be read by the computer, stored in the memory, and executed by the processor.
All references, including publications, patent applications, and patents, cited herein are hereby incorporated by reference to the same extent as if each reference were individually and specifically indicated to be incorporated by reference and were set forth in its entirety herein.
For the purposes of promoting an understanding of the principles of the invention, reference has been made to the preferred embodiments illustrated in the drawings, and specific language has been used to describe these embodiments. However, no limitation of the scope of the invention is intended by this specific language, and the invention should be construed to encompass all embodiments that would normally occur to one of ordinary skill in the art.
The present invention may be described in terms of functional block components and various processing steps. Such functional blocks may be realized by any number of hardware and/or software components configured to perform the specified functions. For example, the present invention may employ various integrated circuit components, e.g., memory elements, processing elements, logic elements, look-up tables, and the like, which may carry out a variety of functions under the control of one or more microprocessors or other control devices. Similarly, where the elements of the present invention are implemented using software programming or software elements the invention may be implemented with any programming or scripting language such as C, C++, Java, assembler, or the like, with the various algorithms being implemented with any combination of data structures, objects, processes, routines or other programming elements. Functional aspects may be implemented in algorithms that execute on one or more processors. Furthermore, the present invention could employ any number of conventional techniques for electronics configuration, signal processing and/or control, data processing and the like. The words “mechanism” and “element” are used broadly and are not limited to mechanical or physical embodiments, but can include software routines in conjunction with processors, etc.
The particular implementations shown and described herein are illustrative examples of the invention and are not intended to otherwise limit the scope of the invention in any way. For the sake of brevity, conventional electronics, control systems, software development and other functional aspects of the systems (and components of the individual operating components of the systems) may not be described in detail. Furthermore, the connecting lines, or connectors shown in the various figures presented are intended to represent exemplary functional relationships and/or physical or logical couplings between the various elements. It should be noted that many alternative or additional functional relationships, physical connections or logical connections may be present in a practical device. Moreover, no item or component is essential to the practice of the invention unless the element is specifically described as “essential” or “critical”.
The use of the terms “a” and “an” and “the” and similar referents in the context of describing the invention (especially in the context of the following claims) are to be construed to cover both the singular and the plural. Furthermore, recitation of ranges of values herein are merely intended to serve as a shorthand method of referring individually to each separate value falling within the range, unless otherwise indicated herein, and each separate value is incorporated into the specification as if it were individually recited herein. Finally, the steps of all methods described herein can be performed in any suitable order unless otherwise indicated herein or otherwise clearly contradicted by context. The use of any and all examples, or exemplary language (e.g., “such as”) provided herein, is intended merely to better illuminate the invention and does not pose a limitation on the scope of the invention unless otherwise claimed. Numerous modifications and adaptations will be readily apparent to those skilled in this art without departing from the spirit and scope of the present invention.

Claims (15)

What is claimed is:
1. A photographing apparatus comprising:
a first imaging device for converting light from a subject received through a first optical system into a first electric signal;
a second imaging device for converting light from the subject received through a second optical system into a second electric signal;
a first image signal processor for generating an image signal for a live view based on the first electric signal output from the first imaging device before a photographing operation of a still image;
an exposure controller for controlling an exposure so as to perform a step exposure in the second imaging device before the photographing of the still image;
an exposure calculator for calculating an exposure amount during the photographing operation of the still image based on the second electric signal converted in the second imaging device obtained through the step exposure;
an exposure amount setting unit for setting an exposure amount of the first or second imaging device based on a highlight reference and an exposure amount of the other imaging devices based on a dark reference at the same time based on the calculated exposure amount; and
a second image processor for generating a first image signal for a still image based on the first electric signal converted by the first imaging device and a second image signal for a still image based on the second electric signal converted by the second imaging device at the same time when the photographing operation of the still image is performed based on the set exposure amount.
2. The photographing apparatus of claim 1, further comprising:
a dislocation detector for detecting a dislocation between the first and second image signals; and
a combining unit for combining the first and second image signals with each other using the detected dislocation.
3. The photographing apparatus of claim 1, further comprising:
a dislocation detector for detecting a dislocation between the first and second image signals; and
a stereo unit for calculating a stereo image parameter.
4. A photographing apparatus comprising:
a first imaging device for converting light from a subject received through a first optical system into a first electric signal;
a second imaging device for converting light from the subject received through a second optical system into a second electric signal;
a first image signal processor for generating an image signal for a live view based on the first electric signal output from the first imaging device before a photographing operation of a still image;
an exposure controller for controlling an exposure so as to perform a step exposure in the second imaging device before the photographing of the still image;
an exposure calculator for calculating an exposure amount during the photographing operation of the still image based on the second electric signal converted in the second imaging device obtained through the step exposure;
an exposure amount setting unit for setting exposure amounts of the first and second imaging devices to be the same as each other at the same timing based on the calculated exposure amount; and
a second image processor for generating a first image signal for a still image based on the first electric signal converted by the first imaging device and a second image signal for a still image based on the second electric signal converted by the second imaging device at the same time when the photographing operation of the still image is performed based on the set exposure amount.
5. The photographing apparatus of claim 4, further comprising:
a dislocation detector for detecting a dislocation between the first and second image signals; and
a combining unit for combining the first and second image signals with each other using the detected dislocation.
6. The photographing apparatus of claim 4, further comprising:
a dislocation detector for detecting a dislocation between the first and second image signals; and
a stereo unit for calculating a stereo image parameter.
7. A photographing method comprising steps of:
converting light from a subject received through a first optical system into a first electric signal by a first imaging device;
converting light from the subject received through a second optical system into a second electric signal by a second imaging device;
generating an image signal for a live view based on the first electric signal converted by the first imaging device before a photographing operation of a still image;
controlling an exposure to be a step exposure in the second imaging device before the photographing operation of the still image;
calculating an exposure amount required in the photographing operation of the still image based on the second electric signal converted by the second imaging device, which is obtained by the step exposure;
setting an exposure amount of the first or second imaging device based on a highlight reference and an exposure amount of the other imaging devices based on a dark reference at the same time based on the calculated exposure amount; and
generating a first image signal for a still image based on the first electric signal converted by the first imaging device and a second image signal for a still image based on the second electric signal converted by the second imaging device at the same time when the photographing operation of the still image is performed based on the set exposure amount.
8. The photographing method of claim 7, further comprising steps of:
detecting a dislocation between the first and second image signals; and
combining the first and second image signals with each other using the detected dislocation.
9. The photographing method of claim 7, further comprising steps of:
detecting a dislocation between the first and second image signals; and
calculating a stereo image parameter.
10. A photographing method comprising steps of:
converting light from a subject received through a first optical system into a first electric signal by a first imaging device;
converting light from the subject received through a second optical system into a second electric signal by a second imaging device;
generating an image signal for a live view based on the first electric signal converted by the first imaging device before a photographing operation of a still image;
controlling an exposure to be a step exposure in the second imaging device before the photographing operation of the still image;
calculating an exposure amount required in the photographing operation of the still image based on the second electric signal converted by the second imaging device, which is obtained by the step exposure;
setting exposure amounts of the first and second imaging devices to be the same as each other at the same timing based on the calculated exposure amount; and
generating a first image signal for a still image based on the first electric signal converted by the first imaging device and a second image signal for a still image based on the second electric signal converted by the second imaging device at the same time when the photographing operation of the still image is performed based on the set exposure amount.
11. The photographing method of claim 10, further comprising steps of:
detecting a dislocation between the first and second image signals; and
combining the first and second image signals with each other using the detected dislocation.
12. The photographing method of claim 10, further comprising steps of:
detecting a dislocation between the first and second image signals; and
calculating a stereo image parameter.
13. A photographing apparatus, comprising:
means for converting light from a subject received through a first optical system into a first electric signal by a first imaging device;
means for converting light from the subject received through a second optical system into a second electric signal by a second imaging device;
means for generating an image signal for a live view based on the first electric signal converted by the first imaging device before a photographing operation of a still image;
means for controlling an exposure to be a step exposure in the second imaging device before the photographing operation of the still image;
means for calculating an exposure amount required in the photographing operation of the still image based on the second electric signal converted by the second imaging device, which is obtained by the step exposure;
means for setting an exposure amount of the first or second imaging device based on a highlight reference and an exposure amount of the other imaging devices based on a dark reference at the same time based on the calculated exposure amount; and
means for generating a first image signal for a still image based on the first electric signal converted by the first imaging device and a second image signal for a still image based on the second electric signal converted by the second imaging device at the same time when the photographing operation of the still image is performed based on the set exposure amount.
14. The photographing apparatus of claim 13, further comprising:
means for detecting a dislocation between the first and second image signals; and
means for combining the first and second image signals with each other using the detected dislocation.
15. The photographing apparatus of claim 13, further comprising:
means for detecting a dislocation between the first and second image signals; and
means for calculating a stereo image parameter.
US12/625,776 2008-11-28 2009-11-25 Photographing apparatus and method for dynamic range adjustment and stereography Active 2032-02-14 US8587691B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/054,077 US8947557B2 (en) 2008-11-28 2013-10-15 Photographing apparatus and method for dynamic range adjustment and stereography

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2008303837A JP5230376B2 (en) 2008-11-28 2008-11-28 Imaging apparatus and imaging method
JP2008-303837 2008-11-28

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US14/054,077 Division US8947557B2 (en) 2008-11-28 2013-10-15 Photographing apparatus and method for dynamic range adjustment and stereography

Publications (2)

Publication Number Publication Date
US20100134652A1 US20100134652A1 (en) 2010-06-03
US8587691B2 true US8587691B2 (en) 2013-11-19

Family

ID=42222485

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/625,776 Active 2032-02-14 US8587691B2 (en) 2008-11-28 2009-11-25 Photographing apparatus and method for dynamic range adjustment and stereography
US14/054,077 Active US8947557B2 (en) 2008-11-28 2013-10-15 Photographing apparatus and method for dynamic range adjustment and stereography

Family Applications After (1)

Application Number Title Priority Date Filing Date
US14/054,077 Active US8947557B2 (en) 2008-11-28 2013-10-15 Photographing apparatus and method for dynamic range adjustment and stereography

Country Status (3)

Country Link
US (2) US8587691B2 (en)
JP (1) JP5230376B2 (en)
KR (1) KR101590871B1 (en)

Cited By (43)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130258055A1 (en) * 2012-03-30 2013-10-03 Altek Corporation Method and device for generating three-dimensional image
US10156706B2 (en) 2014-08-10 2018-12-18 Corephotonics Ltd. Zoom dual-aperture camera with folded lens
US10225479B2 (en) 2013-06-13 2019-03-05 Corephotonics Ltd. Dual aperture zoom digital camera
US10230898B2 (en) 2015-08-13 2019-03-12 Corephotonics Ltd. Dual aperture zoom camera with video support and switching / non-switching dynamic control
US10250797B2 (en) 2013-08-01 2019-04-02 Corephotonics Ltd. Thin multi-aperture imaging system with auto-focus and methods for using same
US10284780B2 (en) 2015-09-06 2019-05-07 Corephotonics Ltd. Auto focus and optical image stabilization with roll compensation in a compact folded camera
US10288897B2 (en) 2015-04-02 2019-05-14 Corephotonics Ltd. Dual voice coil motor structure in a dual-optical module camera
US10288896B2 (en) 2013-07-04 2019-05-14 Corephotonics Ltd. Thin dual-aperture zoom digital camera
US10288840B2 (en) 2015-01-03 2019-05-14 Corephotonics Ltd Miniature telephoto lens module and a camera utilizing such a lens module
US10371928B2 (en) 2015-04-16 2019-08-06 Corephotonics Ltd Auto focus and optical image stabilization in a compact folded camera
US10379371B2 (en) 2015-05-28 2019-08-13 Corephotonics Ltd Bi-directional stiffness for optical image stabilization in a dual-aperture digital camera
US10488631B2 (en) 2016-05-30 2019-11-26 Corephotonics Ltd. Rotational ball-guided voice coil motor
US10534153B2 (en) 2017-02-23 2020-01-14 Corephotonics Ltd. Folded camera lens designs
US10578948B2 (en) 2015-12-29 2020-03-03 Corephotonics Ltd. Dual-aperture zoom digital camera with automatic adjustable tele field of view
US10616484B2 (en) 2016-06-19 2020-04-07 Corephotonics Ltd. Frame syncrhonization in a dual-aperture camera system
US10645286B2 (en) 2017-03-15 2020-05-05 Corephotonics Ltd. Camera with panoramic scanning range
US10694168B2 (en) 2018-04-22 2020-06-23 Corephotonics Ltd. System and method for mitigating or preventing eye damage from structured light IR/NIR projector systems
US10706518B2 (en) 2016-07-07 2020-07-07 Corephotonics Ltd. Dual camera system with improved video smooth transition by image blending
US10845565B2 (en) 2016-07-07 2020-11-24 Corephotonics Ltd. Linear ball guided voice coil motor for folded optic
US10884321B2 (en) 2017-01-12 2021-01-05 Corephotonics Ltd. Compact folded camera
US10904512B2 (en) 2017-09-06 2021-01-26 Corephotonics Ltd. Combined stereoscopic and phase detection depth mapping in a dual aperture camera
USRE48444E1 (en) 2012-11-28 2021-02-16 Corephotonics Ltd. High resolution thin multi-aperture imaging systems
US10951834B2 (en) 2017-10-03 2021-03-16 Corephotonics Ltd. Synthetically enlarged camera aperture
US10976567B2 (en) 2018-02-05 2021-04-13 Corephotonics Ltd. Reduced height penalty for folded camera
US11268830B2 (en) 2018-04-23 2022-03-08 Corephotonics Ltd Optical-path folding-element with an extended two degree of freedom rotation range
US11287081B2 (en) 2019-01-07 2022-03-29 Corephotonics Ltd. Rotation mechanism with sliding joint
US11315276B2 (en) 2019-03-09 2022-04-26 Corephotonics Ltd. System and method for dynamic stereoscopic calibration
US11333955B2 (en) 2017-11-23 2022-05-17 Corephotonics Ltd. Compact folded camera structure
US11363180B2 (en) 2018-08-04 2022-06-14 Corephotonics Ltd. Switchable continuous display information system above camera
US11368631B1 (en) 2019-07-31 2022-06-21 Corephotonics Ltd. System and method for creating background blur in camera panning or motion
US11531209B2 (en) 2016-12-28 2022-12-20 Corephotonics Ltd. Folded camera structure with an extended light-folding-element scanning range
US11637977B2 (en) 2020-07-15 2023-04-25 Corephotonics Ltd. Image sensors and sensing methods to obtain time-of-flight and phase detection information
US11635596B2 (en) 2018-08-22 2023-04-25 Corephotonics Ltd. Two-state zoom folded camera
US11640047B2 (en) 2018-02-12 2023-05-02 Corephotonics Ltd. Folded camera with optical image stabilization
US11659135B2 (en) 2019-10-30 2023-05-23 Corephotonics Ltd. Slow or fast motion video using depth information
US11693064B2 (en) 2020-04-26 2023-07-04 Corephotonics Ltd. Temperature control for Hall bar sensor correction
US11770618B2 (en) 2019-12-09 2023-09-26 Corephotonics Ltd. Systems and methods for obtaining a smart panoramic image
US11770609B2 (en) 2020-05-30 2023-09-26 Corephotonics Ltd. Systems and methods for obtaining a super macro image
US11832018B2 (en) 2020-05-17 2023-11-28 Corephotonics Ltd. Image stitching in the presence of a full field of view reference image
US11910089B2 (en) 2020-07-15 2024-02-20 Corephotonics Lid. Point of view aberrations correction in a scanning folded camera
US11949976B2 (en) 2019-12-09 2024-04-02 Corephotonics Ltd. Systems and methods for obtaining a smart panoramic image
US11946775B2 (en) 2020-07-31 2024-04-02 Corephotonics Ltd. Hall sensor—magnet geometry for large stroke linear position sensing
US11968453B2 (en) 2021-07-22 2024-04-23 Corephotonics Ltd. Optical image stabilization in a scanning folded camera

Families Citing this family (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5283667B2 (en) 2010-07-07 2013-09-04 パナソニック株式会社 Image processing apparatus, image processing method, and program
GB2483433A (en) * 2010-08-31 2012-03-14 Sony Corp 3D image adjustment based on differences in colour properties between stereo pair
US10200671B2 (en) * 2010-12-27 2019-02-05 3Dmedia Corporation Primary and auxiliary image capture devices for image processing and related methods
US8274552B2 (en) * 2010-12-27 2012-09-25 3Dmedia Corporation Primary and auxiliary image capture devices for image processing and related methods
JP5843454B2 (en) * 2011-03-15 2016-01-13 キヤノン株式会社 Image processing apparatus, image processing method, and program
JP5967865B2 (en) 2011-04-01 2016-08-10 キヤノン株式会社 IMAGING DEVICE, IMAGING DEVICE CONTROL METHOD, AND PROGRAM
KR101843450B1 (en) * 2011-08-23 2018-03-29 엘지전자 주식회사 Mobile terminal and method for controlling of the same
US20130076867A1 (en) * 2011-09-28 2013-03-28 Panasonic Corporation Imaging apparatus
JP5860663B2 (en) * 2011-10-18 2016-02-16 日立オートモティブシステムズ株式会社 Stereo imaging device
WO2013133057A1 (en) * 2012-03-07 2013-09-12 ソニー株式会社 Image processing apparatus, method, and program
US9111484B2 (en) * 2012-05-03 2015-08-18 Semiconductor Components Industries, Llc Electronic device for scene evaluation and image projection onto non-planar screens
KR102063102B1 (en) * 2013-08-19 2020-01-07 엘지전자 주식회사 Mobile terminal and control method for the mobile terminal
US9615012B2 (en) * 2013-09-30 2017-04-04 Google Inc. Using a second camera to adjust settings of first camera
US9565416B1 (en) 2013-09-30 2017-02-07 Google Inc. Depth-assisted focus in multi-camera systems
WO2015081563A1 (en) * 2013-12-06 2015-06-11 华为终端有限公司 Method for generating picture and twin-lens device
US10944911B2 (en) 2014-10-24 2021-03-09 Texas Instruments Incorporated Image data processing for digital overlap wide dynamic range sensors
CN106851063A (en) * 2017-02-27 2017-06-13 努比亚技术有限公司 A kind of exposure regulation terminal and method based on dual camera
JP2019197946A (en) * 2018-05-07 2019-11-14 シャープ株式会社 Electronic apparatus, program, control apparatus, and control method

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2003018617A (en) 2001-07-03 2003-01-17 Olympus Optical Co Ltd Imaging apparatus
JP2004135074A (en) 2002-10-10 2004-04-30 Calsonic Kansei Corp Image pickup device
US20050046738A1 (en) 2003-08-25 2005-03-03 Keiji Sato Image sensing apparatus and its control method
US20060140510A1 (en) * 2004-12-27 2006-06-29 Trw Automotive U.S. Llc Method and apparatus for enhancing the dynamic range of stereo vision system
US20070046809A1 (en) 2005-08-29 2007-03-01 Fuji Photo Film Co., Ltd. Image pickup device, multi-eye image pickup device and control program for multi-eye image pickup device
JP2007104408A (en) 2005-10-05 2007-04-19 Ricoh Co Ltd Imaging apparatus and imaging method
JP2007288245A (en) 2006-04-12 2007-11-01 Konica Minolta Photo Imaging Inc Imaging apparatus, image processing method, and image processing program
US20080063294A1 (en) * 2006-09-08 2008-03-13 Peter Jeffrey Burt System and Method for High Performance Image Processing
US20080218612A1 (en) * 2007-03-09 2008-09-11 Border John N Camera using multiple lenses and image sensors in a rangefinder configuration to provide a range map
US20080239129A1 (en) 2007-03-30 2008-10-02 Hiroyuki Oshima Method and device for driving solid-state imaging device, imaging apparatus, and image synthesizing method
US8045792B2 (en) * 2007-03-29 2011-10-25 Samsung Electronics Co., Ltd. Method and apparatus for controlling dynamic depth of stereo-view or multi-view sequence images

Patent Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2003018617A (en) 2001-07-03 2003-01-17 Olympus Optical Co Ltd Imaging apparatus
JP2004135074A (en) 2002-10-10 2004-04-30 Calsonic Kansei Corp Image pickup device
US20050046738A1 (en) 2003-08-25 2005-03-03 Keiji Sato Image sensing apparatus and its control method
JP2005072963A (en) 2003-08-25 2005-03-17 Canon Inc Imaging device and its control method
US20060140510A1 (en) * 2004-12-27 2006-06-29 Trw Automotive U.S. Llc Method and apparatus for enhancing the dynamic range of stereo vision system
US7561731B2 (en) * 2004-12-27 2009-07-14 Trw Automotive U.S. Llc Method and apparatus for enhancing the dynamic range of a stereo vision system
JP2007067574A (en) 2005-08-29 2007-03-15 Fujifilm Corp Imaging apparatus, compound eye imaging apparatus, and control program therefor
US20070046809A1 (en) 2005-08-29 2007-03-01 Fuji Photo Film Co., Ltd. Image pickup device, multi-eye image pickup device and control program for multi-eye image pickup device
JP2007104408A (en) 2005-10-05 2007-04-19 Ricoh Co Ltd Imaging apparatus and imaging method
JP2007288245A (en) 2006-04-12 2007-11-01 Konica Minolta Photo Imaging Inc Imaging apparatus, image processing method, and image processing program
US20080063294A1 (en) * 2006-09-08 2008-03-13 Peter Jeffrey Burt System and Method for High Performance Image Processing
US20080218612A1 (en) * 2007-03-09 2008-09-11 Border John N Camera using multiple lenses and image sensors in a rangefinder configuration to provide a range map
US8045792B2 (en) * 2007-03-29 2011-10-25 Samsung Electronics Co., Ltd. Method and apparatus for controlling dynamic depth of stereo-view or multi-view sequence images
US20080239129A1 (en) 2007-03-30 2008-10-02 Hiroyuki Oshima Method and device for driving solid-state imaging device, imaging apparatus, and image synthesizing method
JP2008252790A (en) 2007-03-30 2008-10-16 Fujifilm Corp Method of driving solid state image pickup device, imaging device, and method of combining image of imaging device

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Notice of Allowance issued for JP 2008-303837 (Feb. 26, 2013).
Office Action established for JP 2008-303837 (Jun. 19, 2012).

Cited By (115)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9258543B2 (en) * 2012-03-30 2016-02-09 Altek Corporation Method and device for generating three-dimensional image
US20130258055A1 (en) * 2012-03-30 2013-10-03 Altek Corporation Method and device for generating three-dimensional image
USRE48697E1 (en) 2012-11-28 2021-08-17 Corephotonics Ltd. High resolution thin multi-aperture imaging systems
USRE49256E1 (en) 2012-11-28 2022-10-18 Corephotonics Ltd. High resolution thin multi-aperture imaging systems
USRE48444E1 (en) 2012-11-28 2021-02-16 Corephotonics Ltd. High resolution thin multi-aperture imaging systems
USRE48477E1 (en) 2012-11-28 2021-03-16 Corephotonics Ltd High resolution thin multi-aperture imaging systems
USRE48945E1 (en) 2012-11-28 2022-02-22 Corephotonics Ltd. High resolution thin multi-aperture imaging systems
US11470257B2 (en) 2013-06-13 2022-10-11 Corephotonics Ltd. Dual aperture zoom digital camera
US11838635B2 (en) 2013-06-13 2023-12-05 Corephotonics Ltd. Dual aperture zoom digital camera
US10904444B2 (en) 2013-06-13 2021-01-26 Corephotonics Ltd. Dual aperture zoom digital camera
US10326942B2 (en) 2013-06-13 2019-06-18 Corephotonics Ltd. Dual aperture zoom digital camera
US10225479B2 (en) 2013-06-13 2019-03-05 Corephotonics Ltd. Dual aperture zoom digital camera
US10841500B2 (en) 2013-06-13 2020-11-17 Corephotonics Ltd. Dual aperture zoom digital camera
US11614635B2 (en) 2013-07-04 2023-03-28 Corephotonics Ltd. Thin dual-aperture zoom digital camera
US10288896B2 (en) 2013-07-04 2019-05-14 Corephotonics Ltd. Thin dual-aperture zoom digital camera
US11287668B2 (en) 2013-07-04 2022-03-29 Corephotonics Ltd. Thin dual-aperture zoom digital camera
US11852845B2 (en) 2013-07-04 2023-12-26 Corephotonics Ltd. Thin dual-aperture zoom digital camera
US10620450B2 (en) 2013-07-04 2020-04-14 Corephotonics Ltd Thin dual-aperture zoom digital camera
US11470235B2 (en) 2013-08-01 2022-10-11 Corephotonics Ltd. Thin multi-aperture imaging system with autofocus and methods for using same
US10469735B2 (en) 2013-08-01 2019-11-05 Corephotonics Ltd. Thin multi-aperture imaging system with auto-focus and methods for using same
US10694094B2 (en) 2013-08-01 2020-06-23 Corephotonics Ltd. Thin multi-aperture imaging system with auto-focus and methods for using same
US11716535B2 (en) 2013-08-01 2023-08-01 Corephotonics Ltd. Thin multi-aperture imaging system with auto-focus and methods for using same
US11856291B2 (en) 2013-08-01 2023-12-26 Corephotonics Ltd. Thin multi-aperture imaging system with auto-focus and methods for using same
US10250797B2 (en) 2013-08-01 2019-04-02 Corephotonics Ltd. Thin multi-aperture imaging system with auto-focus and methods for using same
US10509209B2 (en) 2014-08-10 2019-12-17 Corephotonics Ltd. Zoom dual-aperture camera with folded lens
US11543633B2 (en) 2014-08-10 2023-01-03 Corephotonics Ltd. Zoom dual-aperture camera with folded lens
US10976527B2 (en) 2014-08-10 2021-04-13 Corephotonics Ltd. Zoom dual-aperture camera with folded lens
US10571665B2 (en) 2014-08-10 2020-02-25 Corephotonics Ltd. Zoom dual-aperture camera with folded lens
US11002947B2 (en) 2014-08-10 2021-05-11 Corephotonics Ltd. Zoom dual-aperture camera with folded lens
US11042011B2 (en) 2014-08-10 2021-06-22 Corephotonics Ltd. Zoom dual-aperture camera with folded lens
US11262559B2 (en) 2014-08-10 2022-03-01 Corephotonics Ltd Zoom dual-aperture camera with folded lens
US11703668B2 (en) 2014-08-10 2023-07-18 Corephotonics Ltd. Zoom dual-aperture camera with folded lens
US10156706B2 (en) 2014-08-10 2018-12-18 Corephotonics Ltd. Zoom dual-aperture camera with folded lens
US11125975B2 (en) 2015-01-03 2021-09-21 Corephotonics Ltd. Miniature telephoto lens module and a camera utilizing such a lens module
US10288840B2 (en) 2015-01-03 2019-05-14 Corephotonics Ltd Miniature telephoto lens module and a camera utilizing such a lens module
US10558058B2 (en) 2015-04-02 2020-02-11 Corephontonics Ltd. Dual voice coil motor structure in a dual-optical module camera
US10288897B2 (en) 2015-04-02 2019-05-14 Corephotonics Ltd. Dual voice coil motor structure in a dual-optical module camera
US10656396B1 (en) 2015-04-16 2020-05-19 Corephotonics Ltd. Auto focus and optical image stabilization in a compact folded camera
US10962746B2 (en) 2015-04-16 2021-03-30 Corephotonics Ltd. Auto focus and optical image stabilization in a compact folded camera
US10371928B2 (en) 2015-04-16 2019-08-06 Corephotonics Ltd Auto focus and optical image stabilization in a compact folded camera
US10459205B2 (en) 2015-04-16 2019-10-29 Corephotonics Ltd Auto focus and optical image stabilization in a compact folded camera
US10571666B2 (en) 2015-04-16 2020-02-25 Corephotonics Ltd. Auto focus and optical image stabilization in a compact folded camera
US10613303B2 (en) 2015-04-16 2020-04-07 Corephotonics Ltd. Auto focus and optical image stabilization in a compact folded camera
US11808925B2 (en) 2015-04-16 2023-11-07 Corephotonics Ltd. Auto focus and optical image stabilization in a compact folded camera
US10379371B2 (en) 2015-05-28 2019-08-13 Corephotonics Ltd Bi-directional stiffness for optical image stabilization in a dual-aperture digital camera
US10670879B2 (en) 2015-05-28 2020-06-02 Corephotonics Ltd. Bi-directional stiffness for optical image stabilization in a dual-aperture digital camera
US11770616B2 (en) 2015-08-13 2023-09-26 Corephotonics Ltd. Dual aperture zoom camera with video support and switching / non-switching dynamic control
US11350038B2 (en) 2015-08-13 2022-05-31 Corephotonics Ltd. Dual aperture zoom camera with video support and switching / non-switching dynamic control
US10230898B2 (en) 2015-08-13 2019-03-12 Corephotonics Ltd. Dual aperture zoom camera with video support and switching / non-switching dynamic control
US10917576B2 (en) 2015-08-13 2021-02-09 Corephotonics Ltd. Dual aperture zoom camera with video support and switching / non-switching dynamic control
US10567666B2 (en) 2015-08-13 2020-02-18 Corephotonics Ltd. Dual aperture zoom camera with video support and switching / non-switching dynamic control
US11546518B2 (en) 2015-08-13 2023-01-03 Corephotonics Ltd. Dual aperture zoom camera with video support and switching / non-switching dynamic control
US10356332B2 (en) 2015-08-13 2019-07-16 Corephotonics Ltd. Dual aperture zoom camera with video support and switching / non-switching dynamic control
US10284780B2 (en) 2015-09-06 2019-05-07 Corephotonics Ltd. Auto focus and optical image stabilization with roll compensation in a compact folded camera
US10498961B2 (en) 2015-09-06 2019-12-03 Corephotonics Ltd. Auto focus and optical image stabilization with roll compensation in a compact folded camera
US10578948B2 (en) 2015-12-29 2020-03-03 Corephotonics Ltd. Dual-aperture zoom digital camera with automatic adjustable tele field of view
US11726388B2 (en) 2015-12-29 2023-08-15 Corephotonics Ltd. Dual-aperture zoom digital camera with automatic adjustable tele field of view
US11599007B2 (en) 2015-12-29 2023-03-07 Corephotonics Ltd. Dual-aperture zoom digital camera with automatic adjustable tele field of view
US11392009B2 (en) 2015-12-29 2022-07-19 Corephotonics Ltd. Dual-aperture zoom digital camera with automatic adjustable tele field of view
US10935870B2 (en) 2015-12-29 2021-03-02 Corephotonics Ltd. Dual-aperture zoom digital camera with automatic adjustable tele field of view
US11314146B2 (en) 2015-12-29 2022-04-26 Corephotonics Ltd. Dual-aperture zoom digital camera with automatic adjustable tele field of view
US11650400B2 (en) 2016-05-30 2023-05-16 Corephotonics Ltd. Rotational ball-guided voice coil motor
US10488631B2 (en) 2016-05-30 2019-11-26 Corephotonics Ltd. Rotational ball-guided voice coil motor
US10616484B2 (en) 2016-06-19 2020-04-07 Corephotonics Ltd. Frame syncrhonization in a dual-aperture camera system
US11689803B2 (en) 2016-06-19 2023-06-27 Corephotonics Ltd. Frame synchronization in a dual-aperture camera system
US11172127B2 (en) 2016-06-19 2021-11-09 Corephotonics Ltd. Frame synchronization in a dual-aperture camera system
US11048060B2 (en) 2016-07-07 2021-06-29 Corephotonics Ltd. Linear ball guided voice coil motor for folded optic
US10706518B2 (en) 2016-07-07 2020-07-07 Corephotonics Ltd. Dual camera system with improved video smooth transition by image blending
US10845565B2 (en) 2016-07-07 2020-11-24 Corephotonics Ltd. Linear ball guided voice coil motor for folded optic
US11550119B2 (en) 2016-07-07 2023-01-10 Corephotonics Ltd. Linear ball guided voice coil motor for folded optic
US11531209B2 (en) 2016-12-28 2022-12-20 Corephotonics Ltd. Folded camera structure with an extended light-folding-element scanning range
US11815790B2 (en) 2017-01-12 2023-11-14 Corephotonics Ltd. Compact folded camera
US11809065B2 (en) 2017-01-12 2023-11-07 Corephotonics Ltd. Compact folded camera
US10884321B2 (en) 2017-01-12 2021-01-05 Corephotonics Ltd. Compact folded camera
US11693297B2 (en) 2017-01-12 2023-07-04 Corephotonics Ltd. Compact folded camera
US10670827B2 (en) 2017-02-23 2020-06-02 Corephotonics Ltd. Folded camera lens designs
US10571644B2 (en) 2017-02-23 2020-02-25 Corephotonics Ltd. Folded camera lens designs
US10534153B2 (en) 2017-02-23 2020-01-14 Corephotonics Ltd. Folded camera lens designs
US11671711B2 (en) 2017-03-15 2023-06-06 Corephotonics Ltd. Imaging system with panoramic scanning range
US10645286B2 (en) 2017-03-15 2020-05-05 Corephotonics Ltd. Camera with panoramic scanning range
US10904512B2 (en) 2017-09-06 2021-01-26 Corephotonics Ltd. Combined stereoscopic and phase detection depth mapping in a dual aperture camera
US10951834B2 (en) 2017-10-03 2021-03-16 Corephotonics Ltd. Synthetically enlarged camera aperture
US11695896B2 (en) 2017-10-03 2023-07-04 Corephotonics Ltd. Synthetically enlarged camera aperture
US11809066B2 (en) 2017-11-23 2023-11-07 Corephotonics Ltd. Compact folded camera structure
US11619864B2 (en) 2017-11-23 2023-04-04 Corephotonics Ltd. Compact folded camera structure
US11333955B2 (en) 2017-11-23 2022-05-17 Corephotonics Ltd. Compact folded camera structure
US10976567B2 (en) 2018-02-05 2021-04-13 Corephotonics Ltd. Reduced height penalty for folded camera
US11686952B2 (en) 2018-02-05 2023-06-27 Corephotonics Ltd. Reduced height penalty for folded camera
US11640047B2 (en) 2018-02-12 2023-05-02 Corephotonics Ltd. Folded camera with optical image stabilization
US10694168B2 (en) 2018-04-22 2020-06-23 Corephotonics Ltd. System and method for mitigating or preventing eye damage from structured light IR/NIR projector systems
US10911740B2 (en) 2018-04-22 2021-02-02 Corephotonics Ltd. System and method for mitigating or preventing eye damage from structured light IR/NIR projector systems
US11359937B2 (en) 2018-04-23 2022-06-14 Corephotonics Ltd. Optical-path folding-element with an extended two degree of freedom rotation range
US11268829B2 (en) 2018-04-23 2022-03-08 Corephotonics Ltd Optical-path folding-element with an extended two degree of freedom rotation range
US11733064B1 (en) 2018-04-23 2023-08-22 Corephotonics Ltd. Optical-path folding-element with an extended two degree of freedom rotation range
US11268830B2 (en) 2018-04-23 2022-03-08 Corephotonics Ltd Optical-path folding-element with an extended two degree of freedom rotation range
US11867535B2 (en) 2018-04-23 2024-01-09 Corephotonics Ltd. Optical-path folding-element with an extended two degree of freedom rotation range
US11363180B2 (en) 2018-08-04 2022-06-14 Corephotonics Ltd. Switchable continuous display information system above camera
US11852790B2 (en) 2018-08-22 2023-12-26 Corephotonics Ltd. Two-state zoom folded camera
US11635596B2 (en) 2018-08-22 2023-04-25 Corephotonics Ltd. Two-state zoom folded camera
US11287081B2 (en) 2019-01-07 2022-03-29 Corephotonics Ltd. Rotation mechanism with sliding joint
US11527006B2 (en) 2019-03-09 2022-12-13 Corephotonics Ltd. System and method for dynamic stereoscopic calibration
US11315276B2 (en) 2019-03-09 2022-04-26 Corephotonics Ltd. System and method for dynamic stereoscopic calibration
US11368631B1 (en) 2019-07-31 2022-06-21 Corephotonics Ltd. System and method for creating background blur in camera panning or motion
US11659135B2 (en) 2019-10-30 2023-05-23 Corephotonics Ltd. Slow or fast motion video using depth information
US11770618B2 (en) 2019-12-09 2023-09-26 Corephotonics Ltd. Systems and methods for obtaining a smart panoramic image
US11949976B2 (en) 2019-12-09 2024-04-02 Corephotonics Ltd. Systems and methods for obtaining a smart panoramic image
US11693064B2 (en) 2020-04-26 2023-07-04 Corephotonics Ltd. Temperature control for Hall bar sensor correction
US11832018B2 (en) 2020-05-17 2023-11-28 Corephotonics Ltd. Image stitching in the presence of a full field of view reference image
US11770609B2 (en) 2020-05-30 2023-09-26 Corephotonics Ltd. Systems and methods for obtaining a super macro image
US11962901B2 (en) 2020-05-30 2024-04-16 Corephotonics Ltd. Systems and methods for obtaining a super macro image
US11832008B2 (en) 2020-07-15 2023-11-28 Corephotonics Ltd. Image sensors and sensing methods to obtain time-of-flight and phase detection information
US11637977B2 (en) 2020-07-15 2023-04-25 Corephotonics Ltd. Image sensors and sensing methods to obtain time-of-flight and phase detection information
US11910089B2 (en) 2020-07-15 2024-02-20 Corephotonics Lid. Point of view aberrations correction in a scanning folded camera
US11946775B2 (en) 2020-07-31 2024-04-02 Corephotonics Ltd. Hall sensor—magnet geometry for large stroke linear position sensing
US11968453B2 (en) 2021-07-22 2024-04-23 Corephotonics Ltd. Optical image stabilization in a scanning folded camera

Also Published As

Publication number Publication date
US8947557B2 (en) 2015-02-03
JP5230376B2 (en) 2013-07-10
US20100134652A1 (en) 2010-06-03
KR20100061318A (en) 2010-06-07
JP2010130435A (en) 2010-06-10
US20140036040A1 (en) 2014-02-06
KR101590871B1 (en) 2016-02-02

Similar Documents

Publication Publication Date Title
US8947557B2 (en) Photographing apparatus and method for dynamic range adjustment and stereography
US7706674B2 (en) Device and method for controlling flash
US7796831B2 (en) Digital camera with face detection function for facilitating exposure compensation
KR101605419B1 (en) Photographing apparatus and method of photographing
JP6202860B2 (en) Imaging apparatus and imaging method
US8860840B2 (en) Light source estimation device, light source estimation method, light source estimation program, and imaging apparatus
US8526685B2 (en) Method and apparatus for selectively supporting raw format in digital image processor
JP2010147786A (en) Imaging device and image processing method
JP5156991B2 (en) Imaging apparatus, imaging method, and imaging program
JP2008139683A (en) Imaging apparatus and autofocus control method
JP2018207497A (en) Image processing apparatus and image processing method, imaging apparatus, program, and storage medium
JP2008005540A (en) Image processing apparatus and method
JP2010183460A (en) Image capturing apparatus and method of controlling the same
JP2018050149A (en) Image processing apparatus
JP4170194B2 (en) Imaging device
JP5392336B2 (en) Imaging apparatus, imaging method, and imaging program
JP4871664B2 (en) IMAGING DEVICE AND IMAGING DEVICE CONTROL METHOD
JP2004274344A (en) Image processing apparatus and method
JP2010093780A (en) Imaging apparatus and imaging method
JP5030883B2 (en) Digital still camera and control method thereof
JP2006318260A (en) Image processor and method therefor
JP5091734B2 (en) Imaging apparatus and imaging method
JP2009021893A (en) Imaging device and method
KR101654671B1 (en) Photographing apparatus and photographing method
JP2008283454A (en) Imaging apparatus and imaging method

Legal Events

Date Code Title Description
AS Assignment

Owner name: SAMSUNG DIGITAL IMAGING CO., LTD.,KOREA, REPUBLIC

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TAKANE, YASUO;REEL/FRAME:023883/0323

Effective date: 20091125

Owner name: SAMSUNG DIGITAL IMAGING CO., LTD., KOREA, REPUBLIC

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TAKANE, YASUO;REEL/FRAME:023883/0323

Effective date: 20091125

AS Assignment

Owner name: SAMSUNG ELECTRONICS CO., LTD., KOREA, REPUBLIC OF

Free format text: MERGER;ASSIGNOR:SAMSUNG DIGITAL IMAGING CO., LTD.;REEL/FRAME:026128/0759

Effective date: 20100402

STCF Information on status: patent grant

Free format text: PATENTED CASE

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FPAY Fee payment

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8