US20080259202A1 - Imaging device - Google Patents
Imaging device Download PDFInfo
- Publication number
- US20080259202A1 US20080259202A1 US11/983,962 US98396207A US2008259202A1 US 20080259202 A1 US20080259202 A1 US 20080259202A1 US 98396207 A US98396207 A US 98396207A US 2008259202 A1 US2008259202 A1 US 2008259202A1
- Authority
- US
- United States
- Prior art keywords
- pixels
- imaging
- imaging device
- image
- divided
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000003384 imaging method Methods 0.000 title claims abstract description 206
- 238000012545 processing Methods 0.000 claims abstract description 46
- 238000001514 detection method Methods 0.000 claims abstract description 38
- 230000003287 optical effect Effects 0.000 claims abstract description 28
- 210000001747 pupil Anatomy 0.000 claims description 20
- 230000003321 amplification Effects 0.000 claims description 5
- 238000003199 nucleic acid amplification method Methods 0.000 claims description 5
- 238000011156 evaluation Methods 0.000 description 60
- 238000010586 diagram Methods 0.000 description 27
- 238000012937 correction Methods 0.000 description 14
- 230000006870 function Effects 0.000 description 14
- 230000015654 memory Effects 0.000 description 13
- 230000000875 corresponding effect Effects 0.000 description 11
- 238000004088 simulation Methods 0.000 description 10
- 230000007246 mechanism Effects 0.000 description 7
- 238000004891 communication Methods 0.000 description 6
- 102100035353 Cyclin-dependent kinase 2-associated protein 1 Human genes 0.000 description 5
- 230000008859 change Effects 0.000 description 5
- 238000000034 method Methods 0.000 description 5
- 238000012986 modification Methods 0.000 description 5
- 230000004048 modification Effects 0.000 description 5
- 238000006243 chemical reaction Methods 0.000 description 3
- 239000004973 liquid crystal related substance Substances 0.000 description 3
- 238000012935 Averaging Methods 0.000 description 2
- 239000011248 coating agent Substances 0.000 description 2
- 238000000576 coating method Methods 0.000 description 2
- 230000007423 decrease Effects 0.000 description 2
- 239000003973 paint Substances 0.000 description 2
- 239000000049 pigment Substances 0.000 description 2
- 238000002360 preparation method Methods 0.000 description 2
- 238000001454 recorded image Methods 0.000 description 2
- 230000035945 sensitivity Effects 0.000 description 2
- 101000911772 Homo sapiens Hsc70-interacting protein Proteins 0.000 description 1
- PXHVJJICTQNCMI-UHFFFAOYSA-N Nickel Chemical compound [Ni] PXHVJJICTQNCMI-UHFFFAOYSA-N 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 230000000903 blocking effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000003863 fast low-angle shot imaging Methods 0.000 description 1
- 108090000237 interleukin-24 Proteins 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 229910000652 nickel hydride Inorganic materials 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 238000003825 pressing Methods 0.000 description 1
- 230000011514 reflex Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 230000008054 signal transmission Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/67—Focus control based on electronic image sensor signals
- H04N23/672—Focus control based on electronic image sensor signals based on the phase difference signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/67—Focus control based on electronic image sensor signals
- H04N23/673—Focus control based on electronic image sensor signals based on contrast or high frequency components of image signals, e.g. hill climbing method
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/10—Circuitry of solid-state image sensors [SSIS]; Control thereof for transforming different wavelengths into image signals
- H04N25/11—Arrangement of colour filter arrays [CFA]; Filter mosaics
- H04N25/13—Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements
- H04N25/134—Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements based on three different wavelength filter elements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/70—SSIS architectures; Circuits associated therewith
- H04N25/702—SSIS architectures characterised by non-identical, non-equidistant or non-planar pixel layout
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/70—SSIS architectures; Circuits associated therewith
- H04N25/703—SSIS architectures incorporating pixels for producing signals other than image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/70—SSIS architectures; Circuits associated therewith
- H04N25/703—SSIS architectures incorporating pixels for producing signals other than image signals
- H04N25/704—Pixels specially adapted for focusing, e.g. phase difference pixel sets
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/70—SSIS architectures; Circuits associated therewith
- H04N25/71—Charge-coupled device [CCD] sensors; Charge-transfer registers specially adapted for CCD sensors
- H04N25/75—Circuitry for providing, modifying or processing image signals from the pixel array
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/70—SSIS architectures; Circuits associated therewith
- H04N25/76—Addressed sensors, e.g. MOS or CMOS sensors
- H04N25/78—Readout circuits for addressed sensors, e.g. output amplifiers or A/D converters
Definitions
- the present invention contains subject matter related to Japanese Patent Application JP 2006-319783 filed in the Japan Patent Office on Nov. 28, 2006, the entire contents of which being incorporated herein by reference.
- the present invention relates to an imaging device having an imaging optical system.
- phase difference AF phase difference detection system
- Patent Document 1 discloses also a feature that AF control by a contrast detection system (hereinafter, referred to also as “contrast AF”) is implemented by using outputs from the photoelectric converters in these divided pixels.
- contrast AF AF control by a contrast detection system
- the contrast AF is performed by using the divided pixels provided for the phase difference AF (pixels for phase difference AF). Therefore, there is a fear that the focusing accuracy of this contrast AF is lower than that of existing contrast AF.
- existing contrast AF charge signals of non-divided pixels, of which inside is not divided, are utilized.
- the respective photoelectric converters in divided pixels for receiving light beams that have passed through a part of the lens pupil have sensitivity lower than that of existing non-divided pixels. Accordingly, it is difficult for the contrast AF by use of the outputs of the divided pixels to ensure accuracy equal to that of existing contrast AF.
- an imaging device having an imaging optical system.
- the device includes (a) an imaging element configured to include a plurality of first pixels and a plurality of second pixels arranged along a predetermined direction, (b) a first processor configured to execute focal detection processing by a phase difference detection system based on charge signals obtained from the plurality of second pixels, and (c) a second processor configured to execute specific processing based on charge signals obtained from the plurality of first pixels.
- the specific processing is different from the focal detection processing by a phase difference detection system and is necessary for a function of the imaging device.
- the plurality of first pixels receive a subject light beam that has passed through the entire area of the exit pupil of the imaging optical system, and the plurality of second pixels receive subject light beams that have passed through a pair of partial areas of the exit pupil.
- the first processor creates a pair of image sequences based on charge signals from the second pixels that receive subject light beams that have passed through the pair of partial areas, and detects the amount of shift along the predetermined direction regarding the pair of image sequences, to thereby execute the focal detection processing by a phase difference detection system.
- focal detection processing by a phase difference detection system is executed based on charge signals obtained from the plurality of second pixels that receive subject light beams that have passed through a pair of partial areas of the exit pupil of the imaging optical system. Furthermore, based on charge signals obtained from the plurality of first pixels that receive a subject light beam that has passed through the entire area of the exit pupil of the imaging optical system, specific processing that is different from the focal detection processing by a phase difference detection system and is necessary for a function of the imaging device is executed. As a result, by using an imaging element having pixels (the above-described second pixels) for phase difference AF, specific processing necessary for camera functions other than the phase difference AF can be executed with high accuracy.
- FIG. 1 is a diagram showing the appearance configuration of an imaging device 1 according to an embodiment of the present invention
- FIG. 2 is a diagram showing the appearance configuration of the imaging device 1 ;
- FIG. 3 is a vertical sectional view of the imaging device 1 ;
- FIG. 4 is a block diagram showing the electric configuration of the whole of the imaging device 1 ;
- FIG. 5 is a diagram for explaining the configuration of an imaging element 101 ;
- FIG. 6 is a diagram for explaining the configuration of a G pixel 11 gr having divided inside
- FIG. 7 is a diagram for explaining the principle of phase difference AF employing the imaging element 101 ;
- FIG. 8 is a diagram showing a simulation result when the focal plane is defocused to the 200- ⁇ m-closer side from the imaging plane of the imaging element 101 ;
- FIG. 9 is a diagram showing a simulation result when the focal plane is defocused to the 100- ⁇ m-closer side from the imaging plane;
- FIG. 10 is a diagram showing a simulation result of the focused state in which the focal plane corresponds with the imaging plane
- FIG. 11 is a diagram showing a simulation result when the focal plane is defocused to the 100- ⁇ m-remoter side from the imaging plane;
- FIG. 12 is a diagram showing a simulation result when the focal plane is defocused to the 200- ⁇ m-remoter side from the imaging plane;
- FIG. 13 is a diagram for explaining a graph Gc that indicates the relationship between the defocus amount and the centroid position difference between a pair of image sequences;
- FIG. 14 is a diagram for explaining the principle of contrast AF
- FIG. 15 is a flowchart showing the basic operation of the imaging device 1 ;
- FIG. 16 is a diagram for explaining the AF operation of the imaging device 1 ;
- FIG. 17 is a diagram for explaining the configuration of an imaging element 101 A according to a modification example of the present invention.
- FIG. 18 is a diagram for explaining the configuration of an imaging element 101 B according to another modification example.
- FIGS. 1 and 2 are diagrams showing the appearance configuration of an imaging device 1 according to an embodiment of the present invention.
- FIGS. 1 and 2 are front view and rear view, respectively.
- FIG. 3 is a vertical sectional view of the imaging device 1 .
- the imaging device 1 is configured as e.g. a single-lens reflex digital still camera, and includes a camera body 10 and an imaging lens 2 as an interchangeable lens that can be freely detached from the camera body 10 .
- the following parts are provided on the front face side of the camera body 10 : a mount part 301 that is disposed at substantially the center of the camera front face and on which the imaging lens 2 is mounted; a lens interchange button 302 disposed at the right of the mount part 301 ; a grip part 303 that is provided in a protruding manner at the left end of the camera front face (left side with respect to the X direction) and allows a user to surely grasp the camera with one hand (or both hands); a mode setting dial 305 disposed at upper left part of the camera front face (upper left side with respect to the Y direction); a control value setting dial 306 disposed at upper right part of the camera front face; and a shutter button 307 disposed on the top face of the grip part 303 .
- a liquid crystal display (LCD) 311 a liquid crystal display (LCD) 311 ; a setting button group 312 disposed at the left of the LCD 311 ; an arrow key 314 disposed at the right of the LCD 311 ; and a push button 315 disposed at the center of the arrow key 314 .
- LCD liquid crystal display
- an electronic view finder (EVF) 316 disposed above the LCD 311 ; an eyecup 321 surrounding the EVF 316 ; a main switch 317 disposed at the left of the EVF 316 ; an exposure correction button 323 and an AE lock button 324 that are disposed at the right of the EVF 316 ; and a flash part 318 and a connection terminal 319 that are provided above the EVF 316 .
- EVF electronic view finder
- mount part 301 On the mount part 301 , plural electric contacts for electric connection to the mounted imaging lens 2 , couplers for mechanical connection, and so on are provided.
- the lens interchange button 302 is pushed down at the time of removal of the imaging lens 2 mounted on the mount part 301 .
- the grip part 303 is to allow a user to grasp the imaging device 1 at the time of imaging, and is provided with surface undulation in matching with the finger shape for higher fitting property.
- a cell holder and card holder (not shown) are provided inside the grip part 303 .
- a cell 69 B (see FIG. 4 ) is housed as a camera power source.
- a recording medium (e.g., memory card) for recording the image data of captured images is detachably housed.
- the grip part 303 may be provided with a grip sensor for detecting whether or not the grip part 303 is grasped by a user.
- the mode setting dial 305 and the control value setting dial 306 are formed of a member that has a substantially circular disk shape and can rotate in a plane substantially parallel to the top face of the camera body 10 .
- the mode setting dial 305 is to select one of modes and functions incorporated in the imaging device 1 , such as an auto exposure (AE) control mode, an auto focus (AF) control mode, various imaging modes typified by a still image mode for capturing one still image and a continuous imaging mode for performing continuous imaging, and a reproduction mode for reproducing recorded images.
- the control value setting dial 306 is to set control values for the various functions incorporated in the imaging device 1 .
- the shutter button 307 is a push-down switch that can be halfway pushed down so as to be in the “halfway-pushed state” and can be further pushed down so as to be in the “fully-pushed state”.
- preparation operation for capturing a still image of a subject is executed.
- the shutter button 307 is fully pushed (S 2 )
- imaging operation a series of operation including exposure of an imaging sensor, predetermined image processing for an image signal obtained through the exposure, and recording in a memory card or the like
- the LCD 311 includes a color liquid crystal panel that can display images.
- the LCD 311 is to display an image captured by an imaging element 101 ( FIG. 3 ) and a reproduced image that has been recorded, and is to display a setting screen for the functions and modes incorporated in the imaging device 1 .
- an organic EL display or plasma display may be used.
- the setting button group 312 includes buttons that allow operation for the various functions incorporated in the imaging device 1 .
- the setting button group 312 includes e.g. a selection settlement switch for settling the details selected on a menu screen displayed on the LCD 311 , selection cancel switch, menu displaying switch for a change of the contents of the menu screen, displaying ON/OFF switch, and displaying enlargement switch.
- the arrow key 314 is formed of an annular member that has plural push-down parts (triangle-mark parts in the drawing) disposed along the circumferential direction with a constant interval, and is so configured that pressing operation for the push-down parts is detected through contacts (switches, not shown) provided corresponding to the respective push-down parts.
- the push button 315 is disposed at the center of the arrow key 314 .
- the arrow key 314 and the push button 315 are used to input instructions regarding a change of the imaging magnification (movement of the zoom lens in the wide direction and tele direction), frame stepping of a recorded image to be reproduced on the LCD 311 or the like, setting of imaging conditions (diaphragm value, shutter speed, the presence or absence of flash lighting, etc.), and so on.
- the EVF 316 includes e.g. a color liquid crystal panel that can display images, and is to display an image captured by the imaging element 101 ( FIG. 3 ) and a reproduced image that has been recorded.
- live-view (preview) displaying is performed in which a subject is displayed in a moving image manner based on image signals sequentially created by the imaging element 101 before actual imaging (imaging for image recording). This permits a user to visually recognize the subject to be actually imaged by the imaging element 101 .
- the main switch 317 is formed of a dual-contact slide switch that laterally slides. When the main switch 317 is set to the left, the power supply of the imaging device 1 is turned on. When the main switch 317 is set to the right, the power supply is turned off.
- the flash part 318 is configured as a pop-up built-in flash.
- the connection terminal 319 is used for the connection.
- the eyecup 321 is a U-character shaped light-shielding member that suppresses the entry of external light into the EVF 316 .
- the exposure correction button 323 is to manually adjust the exposure value (diaphragm value and shutter speed), and the AE lock button 324 is to fix exposure.
- the imaging lens 2 functions as a lens window that captures light (optical image) from a subject, and functions also as an imaging optical system for guiding the subject light to the imaging element 101 disposed inside the camera body 10 .
- the imaging lens 2 can be removed from the camera body 10 .
- the imaging lens 2 includes a lens group 21 composed of plural lenses that are serially arranged along an optical axis LT ( FIG. 3 ).
- This lens group 21 includes a focus lens 211 (see FIG. 4 ) for adjustment of the focal point and a zoom lens 212 (see FIG. 4 ) for magnification changes. These lenses are driven in the direction of the optical axis LT, so that the magnification change and focal adjustment are performed. Furthermore, at a proper position on the outer circumference of the lens barrel of the imaging lens 2 , an operating ring that can rotate along the outer circumferential plane of the lens barrel is provided.
- the zoom lens 212 moves in the optical axis direction depending on the rotation direction and rotation amount of the operating ring, so that the zoom magnification (imaging magnification) is set to the value corresponding to the position of the lens movement destination.
- the imaging element 101 is disposed on the optical axis LT of the lens group 21 included in the imaging lens 2 mounted on the camera body 10 in such a manner as to be perpendicular to the optical axis LT.
- CMOS imaging element CMOS imaging element
- plural pixels each having e.g. a photodiode are two-dimensionally arranged in a matrix and e.g. red (R), green (G), and blue (B) color filters having different spectroscopic characteristics are provided at the ratio of 1:2:1 on the light-receiving planes of the respective pixels.
- the imaging element (imaging sensor) 101 converts an optical image of a subject formed through the lens group 21 into analog electric signals (image signals) of the respective color components of R, G, and B, and outputs the signals as R, G, and B image signals.
- image signals analog electric signals
- a shutter unit 40 In front of the imaging element 101 , a shutter unit 40 is disposed.
- This shutter unit 40 has a film body that vertically moves, and is configured as a mechanical focal plane shutter that carries out operations of opening and blocking the optical path of subject light guided to the imaging element 101 along the optical axis LT.
- the shutter unit 40 can be omitted if the imaging element 101 can be fully electronically shuttered.
- FIG. 4 is a block diagram showing the electric configuration of the whole of the imaging device 1 .
- the same members and so on in FIG. 4 as those in FIGS. 1 to 3 are given the same numerals.
- the electric configuration of the imaging lens 2 will be described below.
- the imaging lens 2 includes a lens drive mechanism 24 , a lens position detector 25 , a lens controller 26 , and a diaphragm drive mechanism 27 .
- the focus lens 211 , the zoom lens 212 , and a diaphragm 23 for adjusting the amount of light incident on the imaging element 101 provided in the camera body 10 are held in the lens barrel along the direction of the optical axis LT ( FIG. 3 ). This allows capturing of an optical image of a subject and formation of the optical image on the imaging element 101 .
- the focal adjustment operation is carried out through driving of the lens group 21 in the direction of the optical axis LT by an AF actuator 71 M in the camera body 10 .
- the lens drive mechanism 24 is formed of e.g. a helicoid and a gear (not shown) for rotating the helicoid.
- the lens drive mechanism 24 receives driving force from the AF actuator 71 M via a coupler 74 to thereby move the focus lens 211 and so on in the direction parallel to the optical axis LT.
- the movement direction and movement amount of the focus lens 211 conform to the rotation direction and the number of rotations, respectively, of the AF actuator 71 M.
- the lens position detector 25 includes an encode plate on which plural code patterns are formed along the direction of the optical axis LT with a predetermined pitch within the movement range of the lens group 21 , and an encoder brush that moves integrally with the lens barrel 22 in such a manner as to be in sliding contact with the encode plate.
- the lens position detector 25 detects the movement amount of the lens group 21 at the time of focal adjustment.
- the lens position detected by the lens position detector 25 is output as e.g. the number of pulses.
- the lens controller 26 is formed of e.g. a microcomputer that includes a ROM storing therein a control program and a memory 261 formed of a flash memory or the like storing therein data relating to status information. Furthermore, the lens controller 26 includes a communication unit 262 that communicates with a main controller 62 in the camera body 10 . This communication unit 262 transmits to the main controller 62 e.g. status information data such as the focal length, exit pupil position, diaphragm value, focus distance, and peripheral light amount status of the lens group 21 . On the other hand, the communication unit 262 receives e.g. the drive amount of the focus lens 211 from the main controller 62 .
- data such as focal length information and diaphragm value obtained after the completion of AF operation are transmitted from the communication unit 262 to the main controller 62 .
- the above-described status information data of the lens group 21 data of e.g. the drive amount of the focus lens 211 transmitted from the main controller 62 , and so on are stored.
- the diaphragm drive mechanism 27 receives driving force from a diaphragm drive actuator 73 M via a coupler 75 to thereby change the diaphragm diameter of the diaphragm 23 .
- the camera body 10 includes an analog front-end (AFE) 5 , an image processor 61 , an image memory 614 , the main controller 62 , a flash circuit 63 , an operating unit 64 , VRAMs 65 ( 65 a and 65 b ), a card I/F 66 , a memory card 67 , a communication I/F 68 , a power supply circuit 69 , the cell 69 B, a focus drive controller 71 A, the AF actuator 71 M, a shutter drive controller 72 A, a shutter drive actuator 72 M, a diaphragm drive controller 73 A, and the diaphragm drive actuator 73 M.
- AFE analog front-end
- the imaging element 101 is formed of a CMOS color area sensor as described above.
- a timing control circuit 51 to be described later controls imaging operation such as the start (and end) of exposure operation of the imaging element 101 , selection of the outputs of the respective pixels included in the imaging element 101 , and reading-out of pixel signals.
- the AFE 5 supplies the imaging element 101 with a timing pulse for causing the imaging element 101 to carry out predetermined operation. Furthermore, the AFE 5 executes predetermined signal processing for an image signal (group of analog signals received by the respective pixels of the CMOS area sensor) output from the imaging element 101 , to thereby convert the signal into a digital signal and output it to the image processor 61 .
- This AFE 5 includes the timing control circuit 51 , a signal processor 52 , and an A/D converter 53 .
- the timing control circuit 51 produces predetermined timing pulses (vertical scan pulse ⁇ Vn, horizontal scan pulse ⁇ Vm, and pulses for generating a reset signal ⁇ Vr and so on) based on a reference clock output from the main controller 62 , and outputs the timing pulses to the imaging element 101 for control of the imaging operation of the imaging element 101 .
- the timing control circuit 51 outputs predetermined timing pulses to the signal processor 52 and the A/D converter 53 to thereby control the operation of the signal processor 52 and the A/D converter 53 .
- the signal processor 52 executes predetermined analog signal processing for an analog image signal output from the imaging element 101 .
- This signal processor 52 includes a correlated double sampling (CDS) circuit, an auto gain control (AGC) circuit for amplifying a charge signal output from the imaging element 101 , a clamp circuit, and so on.
- CDS correlated double sampling
- AGC auto gain control
- charge signals from divided G pixels 11 gr to be described later are amplified with a gain (amplification factor) ⁇
- charge signals from non-divided pixels G pixels 11 gb , R pixels 11 r , and B pixels 11 b ) to be described later are amplified with a gain ⁇ different from the gain ⁇ .
- the reason for this amplification with different gains is that the sensitivity of the divided G pixels, which receive light beams that have passed through a part of the exit pupil of the imaging lens 2 , is lower than that of the non-divided pixels, and thus there is a need to amplify signals from the divided G pixels with a gain higher than that for the non-divided pixels to thereby ensure a proper output level.
- the A/D converter 53 converts analog R, G, and B image signals output from the signal processor 52 into a digital image signal composed of plural bits (e.g., 12 bits) based on the timing pulse output from the timing control circuit 51 .
- the image processor 61 executes predetermined signal processing for image data output from the AFE 5 to thereby create an image file, and includes a black level correction circuit 611 , a white balance control circuit 612 , and a gamma correction circuit 613 .
- Image data loaded in the image processor 61 is written to the image memory 614 in synchronization with reading from the imaging element 101 , and from then on the image processor 61 accesses this image data written to the image memory 614 for processing in the respective blocks in the image processor 61 .
- the black level correction circuit 611 corrects the black level of the R, G, and B digital image signals arising from the A/D conversion by the A/D converter 53 to a reference black level.
- the white balance correction circuit 612 carries out level conversion (white balance (WB) adjustment) for the digital signals of the respective color components of R, G, and B based on the reference white dependent upon the light source. Specifically, based on WB adjustment data given from the main controller 62 , the white balance correction circuit 612 specifies from an imaging subject a part that is estimated to be originally white based on luminance, chroma, and other data. Furthermore, the white balance correction circuit 612 calculates, regarding the specified part, the average of the R, G, and B color components, the G/R ratio, and the G/B ratio, and carries out level correction by using the calculated parameters as the correction gains for R and B.
- WB white balance
- the gamma correction circuit 613 corrects the grayscale characteristic of the image data of which WB is adjusted. Specifically, the gamma correction circuit 613 carries out nonlinear transform and offset adjustment by using a gamma correction table in which the level of image data is set in advance for each color component.
- the image memory 614 temporarily stores therein image data output from the image processor 61 , and is used as a work area for predetermined processing by the main controller 62 for this image data. Furthermore, at the time of the reproduction mode, image data read out from the memory card 67 is temporarily stored in the image memory 614 .
- the main controller 62 is formed of e.g. a microcomputer that includes a ROM storing therein a control program and a memory such as a flash memory temporarily storing therein data.
- the main controller 62 controls the operation of the respective units in the imaging device 1 .
- the main controller 62 controls pixel reading of two kinds of modes (live-view read mode and AF read mode).
- the cycle (frame rate) of the pixel reading is set to 60 fps. Furthermore, decimation-reading of pixels is carried out for the imaging element 101 , so that e.g. an image of 640 ⁇ 480 pixels in the VGA size is created as an image for live-view displaying.
- the non-divided pixels to be described later the non-divided G pixels 11 gb , the R pixels 11 r , and the B pixels 11 b
- the images created through the decimation-reading from the imaging element 101 are sequentially displayed on the EVF 316 (or the LCD 311 ), so that live-view displaying of a subject is performed.
- auto focus control is carried out in such a way that the cycle (frame rate) of the pixel reading is set to 240 fps and the divided G pixels 11 gr and the non-divided G pixels 11 gb to be described later are read out. Also in the AF read mode, live-view displaying is performed by reading out at 60 fps the non-divided G pixels 11 gb , the R pixels 11 r , and the B pixels 11 b from the imaging element 101 .
- the flash circuit 63 regulates, in the flash imaging mode, the light emission amount of the flash part 318 or an external flash connected to the connection terminal 319 to the light emission amount designed by the main controller 62 .
- the operating unit 64 includes the above-described mode setting dial 305 , control value setting dial 306 , shutter button 307 , setting button group 312 , arrow key 314 , push button 315 , main switch 317 , and so on.
- the operating unit 64 is used to input operation information to the main controller 62 .
- the VRAMs 65 a and 65 b have memory capacity for image signals corresponding to the numbers of pixels of the LCD 311 and the EVF 316 , and serve as buffer memories between the main controller 62 and the LCD 311 and the EVF 316 .
- the card I/F 66 is an interface for permitting signal transmission/reception between the memory card 67 and the main controller 62 .
- the memory card 67 is a recording medium in which image data created by the main controller 62 is stored.
- the communication I/F 68 is an interface for allowing image data and so on to be transmitted to a personal computer and other external apparatuses.
- the power supply circuit 69 is formed of e.g. a constant voltage circuit, and produces voltage (e.g., 5 V) for driving the entire imaging device 1 , including control units such as the main controller 62 , the imaging element 101 , and other various drive units.
- the energization of the imaging element 101 is controlled based on a control signal supplied from the main controller 62 to the power supply circuit 69 .
- the cell 69 B is a primary cell such as an alkaline dry cell or a secondary cell such as a nickel hydride rechargeable battery, and serves as a power source for supplying the entire imaging device 1 with power.
- the focus drive controller 71 A creates, based on an AF control signal given from the main controller 62 , a drive control signal for the AF actuator 71 M necessary to move the focus lens 211 to the focus position.
- the AF actuator 71 M is formed of a stepping motor or the like, and gives lens driving force to the lens drive mechanism 24 of the imaging lens 2 via the coupler 74 .
- the shutter drive controller 72 A creates a drive control signal for the shutter drive actuator 72 M based on a control signal given from the main controller 62 .
- the shutter drive actuator 72 M drives the shutter unit 40 so that the shutter unit 40 can be opened and closed.
- the diaphragm drive controller 73 A creates a drive control signal for the diaphragm drive actuator 73 M based on a control signal given from the main controller 62 .
- the diaphragm drive actuator 73 M gives driving force to the diaphragm drive mechanism 27 via the coupler 75 .
- the camera body 10 further includes a phase difference AF arithmetic circuit 76 and a contrast AF arithmetic circuit 77 that perform arithmetic operation necessary at the time of auto focus (AF) based on image data of which black level has been corrected, output from the black level correction circuit 611 .
- AF auto focus
- the imaging device 1 is configured to allow AF by a phase difference detection system (phase difference AF), in which the imaging element 101 receives light that has been transmitted (has passed) through different parts of the exit pupil to thereby detect the focal point.
- phase difference AF phase difference detection system
- the configuration of this imaging element 101 and the principle of the phase difference AF employing the imaging element 101 will be described below.
- FIG. 5 is a diagram for explaining the configuration of the imaging element 101 .
- the imaging element 101 has red (R) pixels 11 r , green (G) pixels 11 g , and blue (B) pixels 11 b in which color filters of R, G, and B, respectively, are provided on photodiodes.
- R red
- G green
- B blue
- the G pixels 11 g include plural G pixels 11 gr arranged along the direction of Gr lines L 1 (horizontal direction) and plural G pixels 11 gb arranged along Gb lines L 2 .
- the pixel inside is divided into eight areas along the direction of the Gr lines L 1 .
- eight photoelectric converters 111 to 118 are arranged along the direction of the Gr line L 1 .
- Each of the photoelectric converters 111 to 118 has an independent photodiode, which permits reading out of accumulated charges through photoelectric conversion.
- the charge reading from the imaging element 101 the charge reading from the G pixels 11 gr , of which inside is divided, and that from the other non-divided pixels (the G pixels 11 gb , the R pixels 11 r , and the B pixels 11 b ) can be simultaneously carried out in such a way that the reading method for the G pixels 11 gr is made different from that for the other non-divided pixels.
- the G pixel 11 gr of which inside is divided, will be referred to as “divided G pixel” (referred to also as “divided pixel” simply).
- the G pixel 11 gb of which inside is not divided, will be referred to as “non-divided G pixel” (referred to also as “non-divided pixel” simply).
- phase difference AF by use of the imaging element 101 having the above-described divided G pixels 11 gr will be described in detail below.
- FIG. 7 is a diagram for explaining the principle of the phase difference AF employing the imaging element 101 .
- a light beam Ta that has passed through a right-side part of an exit pupil Ep passes through a green color filter 12 and forms an image on the photoelectric converter 113 , which is the third converter from the left end of the divided G pixel 11 gr .
- a light beam Tb that has passed through a left-side part of the exit pupil Ep passes through the green color filter 12 and forms an image on the photoelectric converter 116 , which is the fifth converter from the left end (third converter from the right end) of the divided G pixel 12 gr .
- the plural non-divided pixels including the non-divided G pixels 11 gb , the R pixels 11 r , and the B pixels 11 b which receive a subject light beam that has passed through the entire area of the exit pupil Ep of the imaging lens 2
- the plural divided G pixels 11 gr receive the subject light beams Ta and Tb that have passed through a pair of partial areas of the exit pupil Ep of the imaging lens 2 .
- light-reception data obtained from the photoelectric converter 113 will be referred to as “A-series data”
- light-reception data obtained from the photoelectric converter 116 will be referred to as “B-series data”.
- FIGS. 8 to 12 showing the A-series data and B-series data obtained from plural divided G pixels 11 gr arranged on one Gr line L 1 ( FIG. 5 ).
- FIG. 8 is a diagram showing a simulation result when the focal plane is defocused to the 200- ⁇ m-closer side from the imaging plane of the imaging element 101 .
- FIG. 9 is a diagram showing a simulation result when the focal plane is defocused to the 100- ⁇ m-closer side from the imaging plane.
- FIG. 10 is a diagram showing a simulation result of the focused state in which the focal plane corresponds with the imaging plane.
- FIG. 11 is a diagram showing a simulation result when the focal plane is defocused to the 100- ⁇ m-remoter side from the imaging plane.
- FIG. 12 is a diagram showing a simulation result when the focal plane is defocused to the 200- ⁇ m-remoter side from the imaging plane.
- the abscissa indicates the positions of the divided G pixels 11 gr with respect to the direction of the Gr line L 1 , while the ordinate indicates outputs from the photoelectric converters 113 and 116 .
- graphs Ga 1 to Ga 5 (represented by full lines) indicate the A-series data
- graphs Gb 1 to Gb 5 (represented by dashed lines) indicate the B-series data.
- the graph Gc shown in FIG. 13 is obtained.
- the abscissa indicates the defocus amount (mm)
- the ordinate indicates the difference in the centroid position (expressed by the number of pixels) between the A-series image sequence and the B-series image sequence.
- the centroid position X g of an image sequence can be obtained in accordance with e.g. Equation (1).
- X g X 1 ⁇ Y 1 + X 2 ⁇ Y 2 + ... + X n ⁇ Y n Y 1 + Y 2 + ... + Y n Equation ⁇ ⁇ ( 1 )
- Equation (1) X 1 to X n denote the pixel positions on the Gr line L 1 from the left end for example, and Y 1 to Y n denote the output values of the pixels at the positions X 1 to X n , respectively.
- Equation (2) the relationship between the defocus amount and the difference in the centroid position between a pair of image sequences is a proportional relationship.
- Equation (2) denotes the slope Gk (represented by the dashed line) of the graph Gc in FIG. 13 , and can be acquired in advance through a factory test or the like.
- the difference in the centroid position (phase difference) regarding the A-series data and B-series data obtained from the divided G pixels 11 gr is obtained in the phase difference AF arithmetic circuit 76 .
- the defocus amount is calculated by using Equation (2), and the drive amount equivalent to the calculated defocus amount is given to the focus lens 211 .
- This allows auto focus (AF) control in which the focus lens 211 is rapidly moved to the detected focal position.
- the relationship between the defocus amount and the drive amount of the focus lens 211 is uniquely determined depending on the design values of the imaging lens 2 mounted on the camera body 10 .
- a pair of image sequences are created based on the respective charge signals from the photoelectric converters 113 and 116 of the divided pixels 11 gr , which receive the subject light beams Ta and Tb that have passed through a pair of partial areas of the exit pupil Ep shown in FIG. 7 . Furthermore, the amount of the error (shift) along the direction of the Gr line L 1 regarding this pair of image sequences is detected to thereby carry out the phase difference AF.
- the imaging device 1 of the present embodiment also employs the contrast AF for highly accurate focusing.
- This contrast AF will be described below.
- a pixel group of the non-divided G pixels 11 gb is read out in the AF area defined in a part (e.g., the center part) of the imaging range, and an AF evaluation value is calculated.
- This AF evaluation value is calculated as e.g. the total sum of the absolute values of the differences between adjacent non-divided G pixels 11 gb in the AF area.
- the relationship like that shown in FIG. 14 is obtained between the respective positions of the focus lens 211 and the AF evaluation values. Specifically, in this relationship, in linkage with the focus lens position change, the AF evaluation value monotonically increases and then monotonically decreases after reaching a peak Qk. The movement of the focus lens 211 is continued until the focal zone is found, i.e., the AF evaluation value passes through the peak Qk.
- the focus position P f of the focus lens 211 can be calculated by using quadratic interpolation approximation expressed by Equation (3).
- contrast AF AF evaluation values are obtained in the contrast AF arithmetic circuit 77 , and the focus lens 211 is moved by the focus drive controller 71 A to the focus position obtained in accordance with Equation (3). This allows auto focus control with high focusing accuracy.
- the imaging device 1 of the present embodiment performs hybrid AF as the combination of the above-described phase difference AF employing the divided G pixels 11 gr and the contrast AF employing the non-divided G pixels 11 gb .
- hybrid AF The specific operation of the imaging device 1 regarding this hybrid AF will be described below.
- FIG. 15 is a flowchart showing the basic operation of the imaging device 1 . This operation is executed by the main controller 62 .
- the power supply to the imaging device 1 is turned on through the main switch 317 , so that the imaging element 101 is activated (step ST 1 ).
- the imaging element 101 is set to the live-view read mode. Specifically, as described above, the frame rate is set to 60 fps. Furthermore, an image relating to live-view displaying is created based on outputs from the non-divided pixels (the non-divided G pixels 11 gb , the R pixels 11 r , and the B pixels 11 b ) of the imaging element 101 , and the created image is displayed on the EVF 316 .
- decimation on a horizontal line basis is carried out in such a way that the data of G pixels on the Gr lines L 1 (divided G pixels 11 gr ), which are absent because being not employed for the image, are interpolated by the data of the non-divided G pixels 11 gb on the Gb lines L 2 , adjacent to the divided G pixels 11 gr in the oblique direction.
- the reason for the creation of a live-view image by use of the non-divided pixels is as follows. Specifically, when an image is created by using the divided G pixels 11 gr , it is difficult to stably create proper images, because the gains of outputs from the photoelectric converters 113 and 116 (see FIG. 7 ) of the divided G pixels 11 gr are adjusted to proper levels by the AGC circuit in the signal processor 52 , and hence there is a possibility that simple addition of these adjusted outputs leads to overflow. Although images with proper levels can be created through averaging of outputs from the photoelectric converters 113 and 116 of the divided G pixels 11 gr , circuits and so on are required for the averaging, which causes complication of the device configuration.
- a live-view image is created in such a way that output signals from the divided pixels are excluded and only output signals from the non-divided pixels are used, in order to stably create proper images without complication of the device configuration.
- a step ST 3 auto exposure control (AE) and auto white balance control (AWB) are implemented. Also in this step, similarly to the above-described processing of creating a live-view image, the AE processing and the AWB processing are executed in such a way that output signals from the divided G pixels 11 gr are not employed but only output signals from the non-divided pixels (the non-divided G pixels 11 gb , the R pixels 11 r , and the B pixels 11 b ), of which inside is not divided, are used.
- a step ST 4 it is determined whether or not the shutter button 307 is halfway pushed (S 1 ) by a user. If the shutter button 307 is halfway pushed, the operation sequence proceeds to a step ST 5 . If it is not halfway pushed, the operation sequence returns to the step ST 2 .
- the imaging element 101 is set to the AF read mode. Specifically, control of the hybrid AF is started with the frame rate set to 240 fps as described above. Also in the AF read mode, live-view displaying is performed based on output signals from the non-divided pixels (the non-divided G pixels 11 gb , the R pixels 11 r , and the B pixels 11 b ), which are read out at 60 fps.
- a step ST 6 based on outputs from the non-divided G pixels 11 gb in the AF area in the imaging element 101 , an AF evaluation value is calculated and acquired by the contrast AF arithmetic circuit 77 . That is, based on an image signal produced in the imaging element 101 , an AF evaluation value (focal detection information) to be used for focal detection by a contrast detection system is acquired.
- a step ST 7 the position of the focus lens 211 is detected based on the number of pulses output from the lens position detector 25 in the imaging lens 2 . It is preferable to detect the position of the focus lens 211 at the intermediate time of the period of exposure of the non-divided pixels used for the calculation of the above-described AF evaluation value.
- phase difference AF is carried out by using output signals from the divided G pixels 11 gr in the imaging element 101 .
- the centroid position difference regarding A-series data and B-series data obtained from the photoelectric converters 113 and 116 of the divided G pixels 11 gr is obtained in the phase difference AF arithmetic circuit 76 , and the focus lens 211 is so driven by using Equation (2) that this centroid position difference will be eliminated. More specifically, the focus lens 211 is driven to the position detected through the phase difference AF (focal detection position).
- a step ST 9 it is determined whether or not the focal adjustment through the phase difference AF has been completed. If the focal adjustment through the phase difference AF has been completed, the operation sequence proceeds to a step ST 10 . If it has not been completed yet, the operation sequence returns to the step ST 6 .
- the steps ST 6 to ST 9 are repeated until the completion of the focal adjustment through the phase difference AF. Due to this repetition, in association with the driving of the focus lens 211 by the phase difference AF, the AF evaluation values (focal detection information) corresponding to the respective positions of the focus lens 211 are acquired as history information of the focal detection. This history information of the AF evaluation value is stored in e.g. a memory in the main controller 62 .
- the final AF evaluation value D m calculated last and the AF evaluation value D m-1 that is previous to the final AF evaluation value D m (previous-to-final AF evaluation value D m-1 ) are acquired.
- a step ST 11 it is determined whether or not the ratio of the final AF evaluation value D m to the previous-to-final AF evaluation value D m-1 is in the range of 0.99 to 1.01 inclusive.
- the purpose of this determination is to regard the position of the focus lens 211 as the focus position to thereby specify the focus position when the ratio of the final AF evaluation value D m to the previous-to-final AF evaluation value D m-1 is in the range of 100% ⁇ 1%, because when the ratio is in this range, the focus lens 211 has been driven to a position corresponding to an AF evaluation value in a range near the peak Qk ( FIG. 14 ), where the slope of the AF evaluation value curve is gentle.
- step ST 11 If it is determined in this step ST 11 that the relationship 0.99 ⁇ D m /D m-1 ⁇ 1.01 is satisfied, i.e., if the focus position of the focus lens 211 is specified based on the above-described history information of the AF evaluation values (the previous-to-final AF evaluation value D m-1 and the final AF evaluation value D m ), the operation sequence proceeds to a step ST 19 . In contrast, if not so, the operation sequence proceeds to a step ST 12 .
- step ST 12 it is determined whether or not the ratio of the final AF evaluation value D m to the previous-to-final AF evaluation value D m-1 is larger than one. If the relationship D m /D m-1 >1 is satisfied, the operation sequence proceeds to a step ST 13 based on a determination that the AF evaluation value is in a monotonic increase state. If the relationship D m /D m-1 ⁇ 1 is satisfied, the operation sequence proceeds to the step ST 19 based on a determination that the AF evaluation value is in a monotonic decrease state.
- an AF evaluation value D 1 is acquired based on outputs from the non-divided G pixels 11 gb in the AF area in the imaging element 101 .
- the above-described final AF evaluation value D m is acquired as the AF evaluation value D 1 from the history information of the AF evaluation value.
- a step ST 14 additional driving of the focus lens 211 is carried out by 1F ⁇ equivalent to the focal depth (depth of field) in the same direction as the drive direction of the phase difference AF.
- F denotes the F number indicating the actual diaphragm regarding the imaging lens (imaging optical system) 2
- a step ST 15 similarly to the step ST 6 , an AF evaluation value D 2 is acquired based on outputs from the non-divided G pixels 11 gb in the AF area in the imaging element 101 .
- a step ST 16 it is determined whether or not the ratio of the AF evaluation value D 2 acquired in the step ST 15 to the AF evaluation value D 1 acquired in the step ST 13 is in the range of 0.99 to 1.01 inclusive.
- the purpose of this determination is as follows. Specifically, when the AF evaluation value D 2 is obtained through additional driving of the focus lens 211 by a drive amount W as shown in FIG. 16 for example in the state in which the AF evaluation value has not reached the peak Qk ( FIG.
- the focus lens 211 has been driven to a position corresponding to an AF evaluation value in a gentle slope area near the peak Qk ( FIG. 14 ). Therefore, through the determination in the step ST 16 , the position of the additionally-driven focus lens 211 is regarded as the focus position to thereby specify the focus position when the AF evaluation value ratio is in the range of 100% ⁇ 1%.
- step ST 16 If it is determined in the step ST 16 that the relationship 0.99 ⁇ D2/D1 ⁇ 1.01 is satisfied, the operation sequence proceeds to the step ST 19 . In contrast, if not so, the operation sequence proceeds to a step ST 17 based on a determination that the focus lens 211 has not reached a position corresponding to an AF evaluation value near the peak Qk ( FIG. 14 ).
- the following processing is executed. Specifically, if the focus position of the focus lens 211 is not specified based on the history information of the AF evaluation values acquired so far, additional driving of the focus lens 211 by a drive amount based on the focal depth relating to the imaging lens 2 is carried out and an AF evaluation value is additionally acquired. Subsequently, the focus position of the focus lens 211 is specified based on the focal detection history information to which this additionally-acquired AF evaluation value (focal detection information) is added.
- the focus lens 211 is back-driven to the position corresponding to the peak of the AF evaluation value. Specifically, if it is determined in the step ST 12 that the ratio of the final AF evaluation value D m to the previous-to-final AF evaluation value D m-1 is lower than or equal to one, it is determined that the focus lens 211 has passed through the position corresponding to the peak of the AF evaluation value, and thus the focus position of the focus lens 211 is specified by using Equation (3). Furthermore, the focus lens 211 , which has passed through the position (focus position) corresponding to the peak of the AF evaluation value through the phase difference AF, is back-driven to the specified focus position.
- the focus lens 211 is driven to this focus position through the operation of the step ST 18 .
- the imaging element 101 is set from the AF read mode to the live-view read mode.
- phase difference AF focal detection processing by a phase difference detection system
- contrast AF different from the phase difference AF, live-view displaying, AE, and AWB are performed based on charge signals obtained from the plural non-divided pixels (first pixels) including the non-divided G pixels 11 gb , the R pixels 11 r , and the B pixels 11 b . Therefore, specific processing necessary for camera functions other than the phase difference AF can be executed with high accuracy.
- the divided pixels 11 gr in which the plural photoelectric converters 111 to 118 are arranged have the photoelectric converters 113 and 116 that create the above-described pair of image sequences (A-series image sequence and B-series image sequence).
- a pair of image sequences used for the phase difference AF can be created easily.
- outputs from the divided pixels and the non-divided pixels are amplified by different gains in the AGC circuit of the signal processor 52 . Therefore, the output level of the divided pixels as well as that of the non-divided pixels can be set to the proper level.
- the respective non-divided pixels 11 gb have a color filter of the same color (green), which allows highly-accurate and proper phase difference AF.
- the specific processing executed based on charge signal obtained from the non-divided pixels encompasses processing of contrast AF, processing relating to auto exposure control (AE), processing relating to auto white balance control (AWB), and processing of creating images relating to live-view displaying (preview displaying).
- contrast AF processing relating to auto exposure control
- ABB auto white balance control
- preview displaying processing of creating images relating to live-view displaying
- phase difference AF it is not essential to carry out phase difference AF by use of an imaging element having divided pixels (divided G pixels), of which inside is divided.
- the phase difference AF may be carried out by using any of imaging elements 101 A and 101 B shown in FIGS. 17 and 18 .
- FIG. 17 is a diagram for explaining the configuration of the imaging element 101 A according to a modification example of the present invention.
- a pair of G pixels 11 g ( 11 gs , 11 gt ) having a green color filter 12 g are arranged to sandwich an R pixel 11 r having a red color filter 12 r .
- a light beam Tb that has passed through a left-side part of an exit pupil Ep passes through the green color filter 12 g and forms an image on a photoelectric converter 110 .
- phase difference AF can be carried out similarly to the above-described divided G pixels 11 gr.
- phase difference AF is carried out by using the G pixels 11 ( 11 gs , 11 gt ) on the Gr lines, while contrast AF, live-view displaying, AE, and AWB are carried out by using R pixels, B pixels, and normal G pixels that are arranged on Gb lines and in which the light-shielding plates 13 a and 13 b are not provided. This allows execution of specific processing necessary for camera functions other than the phase difference AF with high accuracy.
- FIG. 18 is a diagram for explaining the configuration of the imaging element 101 B according to another modification example.
- a pair of G pixels 11 g ( 11 gv , 11 gw ) having a green color filter 12 g are arranged to sandwich an R pixel 11 r having a red color filter 12 r .
- the G pixel 11 gv is provided with a microlens ML of which top surface is provided with a light-shielding layer Qa obtained through black-coating with a pigment or paint for the entire surface other than a light transmissive area Pa equivalent to the slit SLa in FIG. 17 .
- the G pixel 11 gv Due to this microlens ML, in the G pixel 11 gv , a light beam Tb that has passed through a left-side part of an exit pupil Ep passes through the green color filter 12 g and forms an image on a photoelectric converter 110 .
- the G pixel 11 gw is provided with the microlens ML of which top surface is provided with a light-shielding layer Qb obtained through black-coating with a pigment or paint for the entire surface other than a light transmissive area Pb equivalent to the slit SLb in FIG. 17 . Due to this microlens ML, in the G pixel 11 gw , a light beam Ta that has passed through a right-side part of the exit pupil Ep passes through the green color filter 12 g and forms an image on the photoelectric converter 110 .
- phase difference AF can be carried out similarly to the above-described divided G pixels 11 gr.
- phase difference AF is carried out by using the G pixels 11 ( 11 gv , 11 gw ) on the Gr lines
- contrast AF, live-view displaying, AE, and AWB are carried out by using R pixels, B pixels, and normal G pixels that are arranged on Gb lines and in which the light-shielding layers Qa and Qb are not formed on the microlens ML.
- the imaging lens 2 is freely detachable from the camera body 10 .
- the imaging lens 2 may be fixed to the camera body 10 .
- the inside of the G pixel does not necessarily need to be divided into eight areas as long as the pixel inside is divided into two or more areas. Furthermore, it is not essential to divide the G pixels. R pixels or B pixels may be divided.
- the imaging device of the above-described embodiment it is not essential to determine in the step ST 11 of FIG. 15 whether or not the ratio of the final AF evaluation value D m to the previous-to-final AF evaluation value D m-1 is in the range of 100% ⁇ 1%. For example, whether or not the ratio is in the range of 100% ⁇ 3% may be determined.
- the imaging device of the above-described embodiment it is not essential to carry out additional driving of the focus lens 211 by 1F ⁇ in the step ST 14 of FIG. 15 .
- Additional driving by 2F ⁇ may be carried out. That is, the drive amount is based on the focal depth relating to the imaging lens 2 .
- the divided G pixel in the above-described embodiment does not necessarily need to be divided into plural areas along the direction of the Gr lines L 1 shown in FIG. 5 (horizontal direction). It may be divided into plural areas along the vertical direction.
- phase difference AF is carried out based on the amount of the shift in the vertical direction regarding a pair of image sequences (A-series image sequence and B-series image sequence) obtained from the divided G pixels.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- Spectroscopy & Molecular Physics (AREA)
- Studio Devices (AREA)
- Focusing (AREA)
- Automatic Focus Adjustment (AREA)
- Color Television Image Signal Generators (AREA)
Abstract
Herein disclosed is an imaging device having an imaging optical system, the device including: an imaging element configured to include a plurality of first pixels and a plurality of second pixels arranged along a predetermined direction; a first processor configured to execute focal detection processing by a phase difference detection system based on charge signals obtained from the plurality of second pixels; and a second processor configured to execute specific processing based on charge signals obtained from the plurality of first pixels, the specific processing being different from the focal detection processing by a phase difference detection system and being necessary for a function of the imaging device.
Description
- CROSS REFERENCES TO RELATED APPLICATIONS
- The present invention contains subject matter related to Japanese Patent Application JP 2006-319783 filed in the Japan Patent Office on Nov. 28, 2006, the entire contents of which being incorporated herein by reference.
- 1. Field of the Invention
- The present invention relates to an imaging device having an imaging optical system.
- 2. Description of the Related Art
- As a technique in a digital camera (imaging device), a technique is known in which auto focus control by a phase difference detection system (hereinafter, referred to also as “phase difference AF”) is implemented by using an imaging element that includes pixels each having divided inside and plural photoelectric converters (hereinafter, referred to also as “divided pixels”).
- For example, according to the technique disclosed in Japanese Patent Laid-open No. 2001-305415 (hereinafter, Patent Document 1), the respective photoelectric converters in pixels each divided into two areas receive light beams that have passed through different areas of the pupil of an imaging lens (imaging optical system) to thereby create a pair of image sequences, and the shift amount regarding this pair of image sequences is obtained, which allows the phase difference AF with use of an imaging element. Furthermore,
Patent Document 1 discloses also a feature that AF control by a contrast detection system (hereinafter, referred to also as “contrast AF”) is implemented by using outputs from the photoelectric converters in these divided pixels. - However, in the technique of
Patent Document 1, the contrast AF is performed by using the divided pixels provided for the phase difference AF (pixels for phase difference AF). Therefore, there is a fear that the focusing accuracy of this contrast AF is lower than that of existing contrast AF. Specifically, in existing contrast AF, charge signals of non-divided pixels, of which inside is not divided, are utilized. In contrast, the respective photoelectric converters in divided pixels for receiving light beams that have passed through a part of the lens pupil have sensitivity lower than that of existing non-divided pixels. Accordingly, it is difficult for the contrast AF by use of the outputs of the divided pixels to ensure accuracy equal to that of existing contrast AF. - In addition, if auto exposure control (AE), auto white balance control (AWB), and so on are implemented based on the outputs of the divided pixels similarly to the above-described contrast AF, there is a fear that the accuracy of these specific controls necessary for camera functions is also lower than that of existing controls.
- There is a need for the present invention to provide an imaging device that can execute specific processing necessary for camera functions other than phase difference AF with high accuracy by using an imaging element that has pixels for the phase difference AF.
- According to an embodiment of the present invention, there is provided an imaging device having an imaging optical system. The device includes (a) an imaging element configured to include a plurality of first pixels and a plurality of second pixels arranged along a predetermined direction, (b) a first processor configured to execute focal detection processing by a phase difference detection system based on charge signals obtained from the plurality of second pixels, and (c) a second processor configured to execute specific processing based on charge signals obtained from the plurality of first pixels. The specific processing is different from the focal detection processing by a phase difference detection system and is necessary for a function of the imaging device. The plurality of first pixels receive a subject light beam that has passed through the entire area of the exit pupil of the imaging optical system, and the plurality of second pixels receive subject light beams that have passed through a pair of partial areas of the exit pupil. The first processor creates a pair of image sequences based on charge signals from the second pixels that receive subject light beams that have passed through the pair of partial areas, and detects the amount of shift along the predetermined direction regarding the pair of image sequences, to thereby execute the focal detection processing by a phase difference detection system.
- According to the embodiment of the present invention, focal detection processing by a phase difference detection system is executed based on charge signals obtained from the plurality of second pixels that receive subject light beams that have passed through a pair of partial areas of the exit pupil of the imaging optical system. Furthermore, based on charge signals obtained from the plurality of first pixels that receive a subject light beam that has passed through the entire area of the exit pupil of the imaging optical system, specific processing that is different from the focal detection processing by a phase difference detection system and is necessary for a function of the imaging device is executed. As a result, by using an imaging element having pixels (the above-described second pixels) for phase difference AF, specific processing necessary for camera functions other than the phase difference AF can be executed with high accuracy.
-
FIG. 1 is a diagram showing the appearance configuration of animaging device 1 according to an embodiment of the present invention; -
FIG. 2 is a diagram showing the appearance configuration of theimaging device 1; -
FIG. 3 is a vertical sectional view of theimaging device 1; -
FIG. 4 is a block diagram showing the electric configuration of the whole of theimaging device 1; -
FIG. 5 is a diagram for explaining the configuration of animaging element 101; -
FIG. 6 is a diagram for explaining the configuration of aG pixel 11 gr having divided inside; -
FIG. 7 is a diagram for explaining the principle of phase difference AF employing theimaging element 101; -
FIG. 8 is a diagram showing a simulation result when the focal plane is defocused to the 200-μm-closer side from the imaging plane of theimaging element 101; -
FIG. 9 is a diagram showing a simulation result when the focal plane is defocused to the 100-μm-closer side from the imaging plane; -
FIG. 10 is a diagram showing a simulation result of the focused state in which the focal plane corresponds with the imaging plane; -
FIG. 11 is a diagram showing a simulation result when the focal plane is defocused to the 100-μm-remoter side from the imaging plane; -
FIG. 12 is a diagram showing a simulation result when the focal plane is defocused to the 200-μm-remoter side from the imaging plane; -
FIG. 13 is a diagram for explaining a graph Gc that indicates the relationship between the defocus amount and the centroid position difference between a pair of image sequences; -
FIG. 14 is a diagram for explaining the principle of contrast AF; -
FIG. 15 is a flowchart showing the basic operation of theimaging device 1; -
FIG. 16 is a diagram for explaining the AF operation of theimaging device 1; -
FIG. 17 is a diagram for explaining the configuration of an imaging element 101A according to a modification example of the present invention; and -
FIG. 18 is a diagram for explaining the configuration of animaging element 101B according to another modification example. -
FIGS. 1 and 2 are diagrams showing the appearance configuration of animaging device 1 according to an embodiment of the present invention.FIGS. 1 and 2 are front view and rear view, respectively.FIG. 3 is a vertical sectional view of theimaging device 1. - The
imaging device 1 is configured as e.g. a single-lens reflex digital still camera, and includes acamera body 10 and animaging lens 2 as an interchangeable lens that can be freely detached from thecamera body 10. - Referring to
FIG. 1 , the following parts are provided on the front face side of the camera body 10: amount part 301 that is disposed at substantially the center of the camera front face and on which theimaging lens 2 is mounted; alens interchange button 302 disposed at the right of themount part 301; agrip part 303 that is provided in a protruding manner at the left end of the camera front face (left side with respect to the X direction) and allows a user to surely grasp the camera with one hand (or both hands); amode setting dial 305 disposed at upper left part of the camera front face (upper left side with respect to the Y direction); a controlvalue setting dial 306 disposed at upper right part of the camera front face; and ashutter button 307 disposed on the top face of thegrip part 303. - Referring to
FIG. 2 , the following parts are provided on the rear face side of the camera body 10: a liquid crystal display (LCD) 311; asetting button group 312 disposed at the left of theLCD 311; anarrow key 314 disposed at the right of theLCD 311; and apush button 315 disposed at the center of thearrow key 314. Furthermore, the following parts are also provided on the rear face side of the camera body 10: an electronic view finder (EVF) 316 disposed above theLCD 311; aneyecup 321 surrounding theEVF 316; amain switch 317 disposed at the left of theEVF 316; anexposure correction button 323 and anAE lock button 324 that are disposed at the right of theEVF 316; and aflash part 318 and aconnection terminal 319 that are provided above theEVF 316. - On the
mount part 301, plural electric contacts for electric connection to the mountedimaging lens 2, couplers for mechanical connection, and so on are provided. - The
lens interchange button 302 is pushed down at the time of removal of theimaging lens 2 mounted on themount part 301. - The
grip part 303 is to allow a user to grasp theimaging device 1 at the time of imaging, and is provided with surface undulation in matching with the finger shape for higher fitting property. Inside thegrip part 303, a cell holder and card holder (not shown) are provided. In the cell holder, acell 69B (seeFIG. 4 ) is housed as a camera power source. In the card holder, a recording medium (e.g., memory card) for recording the image data of captured images is detachably housed. Thegrip part 303 may be provided with a grip sensor for detecting whether or not thegrip part 303 is grasped by a user. - The
mode setting dial 305 and the controlvalue setting dial 306 are formed of a member that has a substantially circular disk shape and can rotate in a plane substantially parallel to the top face of thecamera body 10. Themode setting dial 305 is to select one of modes and functions incorporated in theimaging device 1, such as an auto exposure (AE) control mode, an auto focus (AF) control mode, various imaging modes typified by a still image mode for capturing one still image and a continuous imaging mode for performing continuous imaging, and a reproduction mode for reproducing recorded images. The controlvalue setting dial 306 is to set control values for the various functions incorporated in theimaging device 1. - The
shutter button 307 is a push-down switch that can be halfway pushed down so as to be in the “halfway-pushed state” and can be further pushed down so as to be in the “fully-pushed state”. When theshutter button 307 is halfway pushed (S1) in a still image mode, preparation operation for capturing a still image of a subject (preparation operation such as setting of the exposure control value and focus adjustment) is executed. When theshutter button 307 is fully pushed (S2), imaging operation (a series of operation including exposure of an imaging sensor, predetermined image processing for an image signal obtained through the exposure, and recording in a memory card or the like) is executed. - The
LCD 311 includes a color liquid crystal panel that can display images. TheLCD 311 is to display an image captured by an imaging element 101 (FIG. 3 ) and a reproduced image that has been recorded, and is to display a setting screen for the functions and modes incorporated in theimaging device 1. Instead of theLCD 311, an organic EL display or plasma display may be used. - The
setting button group 312 includes buttons that allow operation for the various functions incorporated in theimaging device 1. Specifically, thesetting button group 312 includes e.g. a selection settlement switch for settling the details selected on a menu screen displayed on theLCD 311, selection cancel switch, menu displaying switch for a change of the contents of the menu screen, displaying ON/OFF switch, and displaying enlargement switch. - The
arrow key 314 is formed of an annular member that has plural push-down parts (triangle-mark parts in the drawing) disposed along the circumferential direction with a constant interval, and is so configured that pressing operation for the push-down parts is detected through contacts (switches, not shown) provided corresponding to the respective push-down parts. Thepush button 315 is disposed at the center of thearrow key 314. Thearrow key 314 and thepush button 315 are used to input instructions regarding a change of the imaging magnification (movement of the zoom lens in the wide direction and tele direction), frame stepping of a recorded image to be reproduced on theLCD 311 or the like, setting of imaging conditions (diaphragm value, shutter speed, the presence or absence of flash lighting, etc.), and so on. - The
EVF 316 includes e.g. a color liquid crystal panel that can display images, and is to display an image captured by the imaging element 101 (FIG. 3 ) and a reproduced image that has been recorded. On theEVF 316 and theLCD 311, live-view (preview) displaying is performed in which a subject is displayed in a moving image manner based on image signals sequentially created by theimaging element 101 before actual imaging (imaging for image recording). This permits a user to visually recognize the subject to be actually imaged by theimaging element 101. - The
main switch 317 is formed of a dual-contact slide switch that laterally slides. When themain switch 317 is set to the left, the power supply of theimaging device 1 is turned on. When themain switch 317 is set to the right, the power supply is turned off. - The
flash part 318 is configured as a pop-up built-in flash. When an external flash or the like is attached to thecamera body 10, theconnection terminal 319 is used for the connection. - The
eyecup 321 is a U-character shaped light-shielding member that suppresses the entry of external light into theEVF 316. - The
exposure correction button 323 is to manually adjust the exposure value (diaphragm value and shutter speed), and theAE lock button 324 is to fix exposure. - The
imaging lens 2 functions as a lens window that captures light (optical image) from a subject, and functions also as an imaging optical system for guiding the subject light to theimaging element 101 disposed inside thecamera body 10. By pushing down the above-describedlens interchange button 302, theimaging lens 2 can be removed from thecamera body 10. - The
imaging lens 2 includes alens group 21 composed of plural lenses that are serially arranged along an optical axis LT (FIG. 3 ). Thislens group 21 includes a focus lens 211 (seeFIG. 4 ) for adjustment of the focal point and a zoom lens 212 (seeFIG. 4 ) for magnification changes. These lenses are driven in the direction of the optical axis LT, so that the magnification change and focal adjustment are performed. Furthermore, at a proper position on the outer circumference of the lens barrel of theimaging lens 2, an operating ring that can rotate along the outer circumferential plane of the lens barrel is provided. In response to manual or auto operation, the zoom lens 212 moves in the optical axis direction depending on the rotation direction and rotation amount of the operating ring, so that the zoom magnification (imaging magnification) is set to the value corresponding to the position of the lens movement destination. - The
imaging element 101 is disposed on the optical axis LT of thelens group 21 included in theimaging lens 2 mounted on thecamera body 10 in such a manner as to be perpendicular to the optical axis LT. As theimaging element 101, a Bayer-arrangement CMOS color area sensor (CMOS imaging element) is used, in which plural pixels each having e.g. a photodiode are two-dimensionally arranged in a matrix and e.g. red (R), green (G), and blue (B) color filters having different spectroscopic characteristics are provided at the ratio of 1:2:1 on the light-receiving planes of the respective pixels. The imaging element (imaging sensor) 101 converts an optical image of a subject formed through thelens group 21 into analog electric signals (image signals) of the respective color components of R, G, and B, and outputs the signals as R, G, and B image signals. The configuration of thisimaging element 101 will be described in detail later. - In front of the
imaging element 101, ashutter unit 40 is disposed. Thisshutter unit 40 has a film body that vertically moves, and is configured as a mechanical focal plane shutter that carries out operations of opening and blocking the optical path of subject light guided to theimaging element 101 along the optical axis LT. Theshutter unit 40 can be omitted if theimaging element 101 can be fully electronically shuttered. -
FIG. 4 is a block diagram showing the electric configuration of the whole of theimaging device 1. The same members and so on inFIG. 4 as those inFIGS. 1 to 3 are given the same numerals. For convenience of description, initially the electric configuration of theimaging lens 2 will be described below. - In addition to the
lens group 21 serving as the above-described imaging optical system, theimaging lens 2 includes alens drive mechanism 24, alens position detector 25, alens controller 26, and adiaphragm drive mechanism 27. - For the
lens group 21, the focus lens 211, the zoom lens 212, and adiaphragm 23 for adjusting the amount of light incident on theimaging element 101 provided in thecamera body 10 are held in the lens barrel along the direction of the optical axis LT (FIG. 3 ). This allows capturing of an optical image of a subject and formation of the optical image on theimaging element 101. The focal adjustment operation is carried out through driving of thelens group 21 in the direction of the optical axis LT by anAF actuator 71M in thecamera body 10. - The
lens drive mechanism 24 is formed of e.g. a helicoid and a gear (not shown) for rotating the helicoid. Thelens drive mechanism 24 receives driving force from theAF actuator 71M via acoupler 74 to thereby move the focus lens 211 and so on in the direction parallel to the optical axis LT. The movement direction and movement amount of the focus lens 211 conform to the rotation direction and the number of rotations, respectively, of theAF actuator 71M. - The
lens position detector 25 includes an encode plate on which plural code patterns are formed along the direction of the optical axis LT with a predetermined pitch within the movement range of thelens group 21, and an encoder brush that moves integrally with thelens barrel 22 in such a manner as to be in sliding contact with the encode plate. Thelens position detector 25 detects the movement amount of thelens group 21 at the time of focal adjustment. The lens position detected by thelens position detector 25 is output as e.g. the number of pulses. - The
lens controller 26 is formed of e.g. a microcomputer that includes a ROM storing therein a control program and amemory 261 formed of a flash memory or the like storing therein data relating to status information. Furthermore, thelens controller 26 includes acommunication unit 262 that communicates with amain controller 62 in thecamera body 10. Thiscommunication unit 262 transmits to themain controller 62 e.g. status information data such as the focal length, exit pupil position, diaphragm value, focus distance, and peripheral light amount status of thelens group 21. On the other hand, thecommunication unit 262 receives e.g. the drive amount of the focus lens 211 from themain controller 62. Furthermore, at the time of imaging, data such as focal length information and diaphragm value obtained after the completion of AF operation are transmitted from thecommunication unit 262 to themain controller 62. In thememory 261, the above-described status information data of thelens group 21, data of e.g. the drive amount of the focus lens 211 transmitted from themain controller 62, and so on are stored. - The
diaphragm drive mechanism 27 receives driving force from adiaphragm drive actuator 73M via acoupler 75 to thereby change the diaphragm diameter of thediaphragm 23. - The electric configuration of the
camera body 10 will be described below. In addition to the above-describedimaging element 101,shutter unit 40, and so on, thecamera body 10 includes an analog front-end (AFE) 5, animage processor 61, animage memory 614, themain controller 62, aflash circuit 63, an operatingunit 64, VRAMs 65 (65 a and 65 b), a card I/F 66, amemory card 67, a communication I/F 68, apower supply circuit 69, thecell 69B, a focus drive controller 71A, theAF actuator 71M, ashutter drive controller 72A, ashutter drive actuator 72M, adiaphragm drive controller 73 A, and thediaphragm drive actuator 73M. - The
imaging element 101 is formed of a CMOS color area sensor as described above. Atiming control circuit 51 to be described later controls imaging operation such as the start (and end) of exposure operation of theimaging element 101, selection of the outputs of the respective pixels included in theimaging element 101, and reading-out of pixel signals. - The
AFE 5 supplies theimaging element 101 with a timing pulse for causing theimaging element 101 to carry out predetermined operation. Furthermore, theAFE 5 executes predetermined signal processing for an image signal (group of analog signals received by the respective pixels of the CMOS area sensor) output from theimaging element 101, to thereby convert the signal into a digital signal and output it to theimage processor 61. ThisAFE 5 includes thetiming control circuit 51, asignal processor 52, and an A/D converter 53. - The
timing control circuit 51 produces predetermined timing pulses (vertical scan pulse φVn, horizontal scan pulse φVm, and pulses for generating a reset signal φVr and so on) based on a reference clock output from themain controller 62, and outputs the timing pulses to theimaging element 101 for control of the imaging operation of theimaging element 101. In addition, thetiming control circuit 51 outputs predetermined timing pulses to thesignal processor 52 and the A/D converter 53 to thereby control the operation of thesignal processor 52 and the A/D converter 53. - The
signal processor 52 executes predetermined analog signal processing for an analog image signal output from theimaging element 101. Thissignal processor 52 includes a correlated double sampling (CDS) circuit, an auto gain control (AGC) circuit for amplifying a charge signal output from theimaging element 101, a clamp circuit, and so on. - In the AGC circuit of the
signal processor 52, charge signals from dividedG pixels 11 gr to be described later are amplified with a gain (amplification factor) α, and charge signals from non-divided pixels (G pixels 11 gb,R pixels 11 r, andB pixels 11 b) to be described later are amplified with a gain β different from the gain α. The reason for this amplification with different gains is that the sensitivity of the divided G pixels, which receive light beams that have passed through a part of the exit pupil of theimaging lens 2, is lower than that of the non-divided pixels, and thus there is a need to amplify signals from the divided G pixels with a gain higher than that for the non-divided pixels to thereby ensure a proper output level. - The A/
D converter 53 converts analog R, G, and B image signals output from thesignal processor 52 into a digital image signal composed of plural bits (e.g., 12 bits) based on the timing pulse output from thetiming control circuit 51. - The
image processor 61 executes predetermined signal processing for image data output from theAFE 5 to thereby create an image file, and includes a blacklevel correction circuit 611, a whitebalance control circuit 612, and agamma correction circuit 613. Image data loaded in theimage processor 61 is written to theimage memory 614 in synchronization with reading from theimaging element 101, and from then on theimage processor 61 accesses this image data written to theimage memory 614 for processing in the respective blocks in theimage processor 61. - The black
level correction circuit 611 corrects the black level of the R, G, and B digital image signals arising from the A/D conversion by the A/D converter 53 to a reference black level. - The white
balance correction circuit 612 carries out level conversion (white balance (WB) adjustment) for the digital signals of the respective color components of R, G, and B based on the reference white dependent upon the light source. Specifically, based on WB adjustment data given from themain controller 62, the whitebalance correction circuit 612 specifies from an imaging subject a part that is estimated to be originally white based on luminance, chroma, and other data. Furthermore, the whitebalance correction circuit 612 calculates, regarding the specified part, the average of the R, G, and B color components, the G/R ratio, and the G/B ratio, and carries out level correction by using the calculated parameters as the correction gains for R and B. - The
gamma correction circuit 613 corrects the grayscale characteristic of the image data of which WB is adjusted. Specifically, thegamma correction circuit 613 carries out nonlinear transform and offset adjustment by using a gamma correction table in which the level of image data is set in advance for each color component. - At the time of the imaging mode, the
image memory 614 temporarily stores therein image data output from theimage processor 61, and is used as a work area for predetermined processing by themain controller 62 for this image data. Furthermore, at the time of the reproduction mode, image data read out from thememory card 67 is temporarily stored in theimage memory 614. - The
main controller 62 is formed of e.g. a microcomputer that includes a ROM storing therein a control program and a memory such as a flash memory temporarily storing therein data. Themain controller 62 controls the operation of the respective units in theimaging device 1. - Furthermore, for the
imaging element 101, themain controller 62 controls pixel reading of two kinds of modes (live-view read mode and AF read mode). - In the live-view read mode of the
imaging element 101, the cycle (frame rate) of the pixel reading is set to 60 fps. Furthermore, decimation-reading of pixels is carried out for theimaging element 101, so that e.g. an image of 640×480 pixels in the VGA size is created as an image for live-view displaying. In this decimation-reading, the non-divided pixels to be described later (thenon-divided G pixels 11 gb, theR pixels 11 r, and theB pixels 11 b) are read out. The images created through the decimation-reading from theimaging element 101 are sequentially displayed on the EVF 316 (or the LCD 311), so that live-view displaying of a subject is performed. - In the AF read mode of the
imaging element 101, auto focus control is carried out in such a way that the cycle (frame rate) of the pixel reading is set to 240 fps and the dividedG pixels 11 gr and thenon-divided G pixels 11 gb to be described later are read out. Also in the AF read mode, live-view displaying is performed by reading out at 60 fps thenon-divided G pixels 11 gb, theR pixels 11 r, and theB pixels 11 b from theimaging element 101. - The
flash circuit 63 regulates, in the flash imaging mode, the light emission amount of theflash part 318 or an external flash connected to theconnection terminal 319 to the light emission amount designed by themain controller 62. - The operating
unit 64 includes the above-describedmode setting dial 305, controlvalue setting dial 306,shutter button 307, settingbutton group 312,arrow key 314,push button 315,main switch 317, and so on. The operatingunit 64 is used to input operation information to themain controller 62. - The
VRAMs LCD 311 and theEVF 316, and serve as buffer memories between themain controller 62 and theLCD 311 and theEVF 316. The card I/F 66 is an interface for permitting signal transmission/reception between thememory card 67 and themain controller 62. Thememory card 67 is a recording medium in which image data created by themain controller 62 is stored. The communication I/F 68 is an interface for allowing image data and so on to be transmitted to a personal computer and other external apparatuses. - The
power supply circuit 69 is formed of e.g. a constant voltage circuit, and produces voltage (e.g., 5 V) for driving theentire imaging device 1, including control units such as themain controller 62, theimaging element 101, and other various drive units. The energization of theimaging element 101 is controlled based on a control signal supplied from themain controller 62 to thepower supply circuit 69. Thecell 69B is a primary cell such as an alkaline dry cell or a secondary cell such as a nickel hydride rechargeable battery, and serves as a power source for supplying theentire imaging device 1 with power. - The focus drive controller 71A creates, based on an AF control signal given from the
main controller 62, a drive control signal for theAF actuator 71M necessary to move the focus lens 211 to the focus position. TheAF actuator 71M is formed of a stepping motor or the like, and gives lens driving force to thelens drive mechanism 24 of theimaging lens 2 via thecoupler 74. - The
shutter drive controller 72A creates a drive control signal for theshutter drive actuator 72M based on a control signal given from themain controller 62. Theshutter drive actuator 72M drives theshutter unit 40 so that theshutter unit 40 can be opened and closed. - The
diaphragm drive controller 73A creates a drive control signal for thediaphragm drive actuator 73M based on a control signal given from themain controller 62. Thediaphragm drive actuator 73M gives driving force to thediaphragm drive mechanism 27 via thecoupler 75. - The
camera body 10 further includes a phase difference AFarithmetic circuit 76 and a contrast AFarithmetic circuit 77 that perform arithmetic operation necessary at the time of auto focus (AF) based on image data of which black level has been corrected, output from the blacklevel correction circuit 611. - A detailed description will be made below about AF operation of the
imaging device 1 by use of the phase difference AFarithmetic circuit 76 and the contrast AFarithmetic circuit 77. - The
imaging device 1 is configured to allow AF by a phase difference detection system (phase difference AF), in which theimaging element 101 receives light that has been transmitted (has passed) through different parts of the exit pupil to thereby detect the focal point. The configuration of thisimaging element 101 and the principle of the phase difference AF employing theimaging element 101 will be described below. -
FIG. 5 is a diagram for explaining the configuration of theimaging element 101. - The
imaging element 101 has red (R)pixels 11 r, green (G)pixels 11 g, and blue (B)pixels 11 b in which color filters of R, G, and B, respectively, are provided on photodiodes. For each of the pixels 11 (11 r, 11 g, 11 b), one microlens ML is provided. For convenience of illustration, adjacent microlenses ML overlap with each other inFIG. 5 . However, in a practical imaging element, the microlenses ML are arranged without overlapping. - The
G pixels 11 g includeplural G pixels 11 gr arranged along the direction of Gr lines L1 (horizontal direction) andplural G pixels 11 gb arranged along Gb lines L2. In each of theG pixels 11 gr on the Gr lines L1, the pixel inside is divided into eight areas along the direction of the Gr lines L1. Specifically, as shown inFIG. 6 , in theG pixel 11 gr, eightphotoelectric converters 111 to 118 are arranged along the direction of the Gr line L1. Each of thephotoelectric converters 111 to 118 has an independent photodiode, which permits reading out of accumulated charges through photoelectric conversion. In the charge reading from theimaging element 101, the charge reading from theG pixels 11 gr, of which inside is divided, and that from the other non-divided pixels (theG pixels 11 gb, theR pixels 11 r, and theB pixels 11 b) can be simultaneously carried out in such a way that the reading method for theG pixels 11 gr is made different from that for the other non-divided pixels. Hereinafter, theG pixel 11 gr, of which inside is divided, will be referred to as “divided G pixel” (referred to also as “divided pixel” simply). On the other hand, theG pixel 11 gb, of which inside is not divided, will be referred to as “non-divided G pixel” (referred to also as “non-divided pixel” simply). - The principle of the phase difference AF by use of the
imaging element 101 having the above-described dividedG pixels 11 gr will be described in detail below. -
FIG. 7 is a diagram for explaining the principle of the phase difference AF employing theimaging element 101. - The description of the principle is based on the following assumption. Specifically, when the actual diaphragm of the imaging lens (imaging optical system) 2 is equivalent to e.g. F5.6, in the divided
G pixel 11 gr, a light beam Ta that has passed through a right-side part of an exit pupil Ep passes through agreen color filter 12 and forms an image on thephotoelectric converter 113, which is the third converter from the left end of the dividedG pixel 11 gr. On the other hand, a light beam Tb that has passed through a left-side part of the exit pupil Ep passes through thegreen color filter 12 and forms an image on thephotoelectric converter 116, which is the fifth converter from the left end (third converter from the right end) of the dividedG pixel 12 gr. That is, in contrast to the plural non-divided pixels including thenon-divided G pixels 11 gb, theR pixels 11 r, and theB pixels 11 b, which receive a subject light beam that has passed through the entire area of the exit pupil Ep of theimaging lens 2, the plural dividedG pixels 11 gr receive the subject light beams Ta and Tb that have passed through a pair of partial areas of the exit pupil Ep of theimaging lens 2. Hereinafter, light-reception data obtained from thephotoelectric converter 113 will be referred to as “A-series data”, while light-reception data obtained from thephotoelectric converter 116 will be referred to as “B-series data”. In the following, the principle of the phase difference AF will be described with reference toFIGS. 8 to 12 showing the A-series data and B-series data obtained from plural dividedG pixels 11 gr arranged on one Gr line L1 (FIG. 5 ). -
FIG. 8 is a diagram showing a simulation result when the focal plane is defocused to the 200-μm-closer side from the imaging plane of theimaging element 101.FIG. 9 is a diagram showing a simulation result when the focal plane is defocused to the 100-μm-closer side from the imaging plane.FIG. 10 is a diagram showing a simulation result of the focused state in which the focal plane corresponds with the imaging plane.FIG. 11 is a diagram showing a simulation result when the focal plane is defocused to the 100-μm-remoter side from the imaging plane.FIG. 12 is a diagram showing a simulation result when the focal plane is defocused to the 200-μm-remoter side from the imaging plane. InFIGS. 8 to 12 , the abscissa indicates the positions of the dividedG pixels 11 gr with respect to the direction of the Gr line L1, while the ordinate indicates outputs from thephotoelectric converters FIGS. 8 to 12 , graphs Ga1 to Ga5 (represented by full lines) indicate the A-series data, while graphs Gb1 to Gb5 (represented by dashed lines) indicate the B-series data. - Comparison between the respective A-series image sequences and the respective B-series image sequences, which are indicated by the A-series graphs Ga1 to Ga5 and the B-series graphs Gb1 to Gb5 in
FIGS. 8 to 12 , makes it apparent that a larger defocus amount results in a larger amount of the shift (error) along the direction of the Gr line L1 between the A-series image sequence and the B-series image sequence. - When the relationship between the defocus amount and the amount of the shift between a pair of image sequences (A-series and B-series image sequences) is translated into a graph, the graph Gc shown in
FIG. 13 is obtained. InFIG. 13 , the abscissa indicates the defocus amount (mm), while the ordinate indicates the difference in the centroid position (expressed by the number of pixels) between the A-series image sequence and the B-series image sequence. The centroid position Xg of an image sequence can be obtained in accordance with e.g. Equation (1). -
- In Equation (1), X1 to Xn denote the pixel positions on the Gr line L1 from the left end for example, and Y1 to Yn denote the output values of the pixels at the positions X1 to Xn, respectively.
- As shown by the graph Gc in
FIG. 13 , the relationship between the defocus amount and the difference in the centroid position between a pair of image sequences is a proportional relationship. When this relationship is expressed as an equation in which the defocus amount is DF (μm) and the centroid position difference is C (μm), Equation (2) is obtained. -
DF=k×C Equation (2) - The coefficient k in Equation (2) denotes the slope Gk (represented by the dashed line) of the graph Gc in
FIG. 13 , and can be acquired in advance through a factory test or the like. - As described above, the difference in the centroid position (phase difference) regarding the A-series data and B-series data obtained from the divided
G pixels 11 gr is obtained in the phase difference AFarithmetic circuit 76. Subsequently, the defocus amount is calculated by using Equation (2), and the drive amount equivalent to the calculated defocus amount is given to the focus lens 211. This allows auto focus (AF) control in which the focus lens 211 is rapidly moved to the detected focal position. The relationship between the defocus amount and the drive amount of the focus lens 211 is uniquely determined depending on the design values of theimaging lens 2 mounted on thecamera body 10. - That is, in the
imaging device 1, a pair of image sequences are created based on the respective charge signals from thephotoelectric converters pixels 11 gr, which receive the subject light beams Ta and Tb that have passed through a pair of partial areas of the exit pupil Ep shown inFIG. 7 . Furthermore, the amount of the error (shift) along the direction of the Gr line L1 regarding this pair of image sequences is detected to thereby carry out the phase difference AF. - In view of the depth of field of a typical digital camera, it is preferable to carry out final focusing through focal detection by a contrast detection system (contrast AF), of which focusing accuracy is higher than that of the phase difference AF. Therefore, the
imaging device 1 of the present embodiment also employs the contrast AF for highly accurate focusing. The principle of this contrast AF will be described below. - In the contrast AF in the
imaging device 1, a pixel group of thenon-divided G pixels 11 gb is read out in the AF area defined in a part (e.g., the center part) of the imaging range, and an AF evaluation value is calculated. This AF evaluation value is calculated as e.g. the total sum of the absolute values of the differences between adjacentnon-divided G pixels 11 gb in the AF area. - If the AF evaluation values are sequentially calculated in linkage with the movement of the focus lens 211 in a constant direction, the relationship like that shown in
FIG. 14 is obtained between the respective positions of the focus lens 211 and the AF evaluation values. Specifically, in this relationship, in linkage with the focus lens position change, the AF evaluation value monotonically increases and then monotonically decreases after reaching a peak Qk. The movement of the focus lens 211 is continued until the focal zone is found, i.e., the AF evaluation value passes through the peak Qk. - If AF evaluation values Dn-1, Dn, and Dn+1, near the peak Qk and the corresponding points Pn-1, Pn, and Pn+1 of the focus lens 211 are acquired as shown in
FIG. 14 , the focus position Pf of the focus lens 211 can be calculated by using quadratic interpolation approximation expressed by Equation (3). -
- In the contrast AF, AF evaluation values are obtained in the contrast AF
arithmetic circuit 77, and the focus lens 211 is moved by the focus drive controller 71A to the focus position obtained in accordance with Equation (3). This allows auto focus control with high focusing accuracy. - The
imaging device 1 of the present embodiment performs hybrid AF as the combination of the above-described phase difference AF employing the dividedG pixels 11 gr and the contrast AF employing thenon-divided G pixels 11 gb. The specific operation of theimaging device 1 regarding this hybrid AF will be described below. - The above description has dealt with mountain-climbing AF based on a general contrast system. However, the present embodiment allows movement to the focus position through the hybrid AF to be described below even when the AF evaluation value does not pass through the peak Qk.
-
FIG. 15 is a flowchart showing the basic operation of theimaging device 1. This operation is executed by themain controller 62. - Initially, the power supply to the
imaging device 1 is turned on through themain switch 317, so that theimaging element 101 is activated (step ST1). - In a
step ST 2, theimaging element 101 is set to the live-view read mode. Specifically, as described above, the frame rate is set to 60 fps. Furthermore, an image relating to live-view displaying is created based on outputs from the non-divided pixels (thenon-divided G pixels 11 gb, theR pixels 11 r, and theB pixels 11 b) of theimaging element 101, and the created image is displayed on theEVF 316. In the creation of the image, decimation on a horizontal line basis is carried out in such a way that the data of G pixels on the Gr lines L1 (dividedG pixels 11 gr), which are absent because being not employed for the image, are interpolated by the data of thenon-divided G pixels 11 gb on the Gb lines L2, adjacent to the dividedG pixels 11 gr in the oblique direction. - The reason for the creation of a live-view image by use of the non-divided pixels (the
non-divided G pixels 11 gb, theR pixels 11 r, and theB pixels 11 b) is as follows. Specifically, when an image is created by using the dividedG pixels 11 gr, it is difficult to stably create proper images, because the gains of outputs from thephotoelectric converters 113 and 116 (seeFIG. 7 ) of the dividedG pixels 11 gr are adjusted to proper levels by the AGC circuit in thesignal processor 52, and hence there is a possibility that simple addition of these adjusted outputs leads to overflow. Although images with proper levels can be created through averaging of outputs from thephotoelectric converters G pixels 11 gr, circuits and so on are required for the averaging, which causes complication of the device configuration. - Therefore, in the
imaging device 1 of the present embodiment, a live-view image is created in such a way that output signals from the divided pixels are excluded and only output signals from the non-divided pixels are used, in order to stably create proper images without complication of the device configuration. - In a step ST3, auto exposure control (AE) and auto white balance control (AWB) are implemented. Also in this step, similarly to the above-described processing of creating a live-view image, the AE processing and the AWB processing are executed in such a way that output signals from the divided
G pixels 11 gr are not employed but only output signals from the non-divided pixels (thenon-divided G pixels 11 gb, theR pixels 11 r, and theB pixels 11 b), of which inside is not divided, are used. - In a step ST4, it is determined whether or not the
shutter button 307 is halfway pushed (S1) by a user. If theshutter button 307 is halfway pushed, the operation sequence proceeds to a step ST5. If it is not halfway pushed, the operation sequence returns to the step ST2. - In the step ST5, the
imaging element 101 is set to the AF read mode. Specifically, control of the hybrid AF is started with the frame rate set to 240 fps as described above. Also in the AF read mode, live-view displaying is performed based on output signals from the non-divided pixels (thenon-divided G pixels 11 gb, theR pixels 11 r, and theB pixels 11 b), which are read out at 60 fps. - In a step ST6, based on outputs from the
non-divided G pixels 11 gb in the AF area in theimaging element 101, an AF evaluation value is calculated and acquired by the contrast AFarithmetic circuit 77. That is, based on an image signal produced in theimaging element 101, an AF evaluation value (focal detection information) to be used for focal detection by a contrast detection system is acquired. - In a step ST7, the position of the focus lens 211 is detected based on the number of pulses output from the
lens position detector 25 in theimaging lens 2. It is preferable to detect the position of the focus lens 211 at the intermediate time of the period of exposure of the non-divided pixels used for the calculation of the above-described AF evaluation value. - In a step ST8, phase difference AF is carried out by using output signals from the divided
G pixels 11 gr in theimaging element 101. Specifically, the centroid position difference regarding A-series data and B-series data obtained from thephotoelectric converters G pixels 11 gr is obtained in the phase difference AFarithmetic circuit 76, and the focus lens 211 is so driven by using Equation (2) that this centroid position difference will be eliminated. More specifically, the focus lens 211 is driven to the position detected through the phase difference AF (focal detection position). - In a step ST9, it is determined whether or not the focal adjustment through the phase difference AF has been completed. If the focal adjustment through the phase difference AF has been completed, the operation sequence proceeds to a step ST10. If it has not been completed yet, the operation sequence returns to the step ST6.
- The steps ST6 to ST9 are repeated until the completion of the focal adjustment through the phase difference AF. Due to this repetition, in association with the driving of the focus lens 211 by the phase difference AF, the AF evaluation values (focal detection information) corresponding to the respective positions of the focus lens 211 are acquired as history information of the focal detection. This history information of the AF evaluation value is stored in e.g. a memory in the
main controller 62. - In the step ST10, of the history information of the AF evaluation values calculated during the phase difference AF, the final AF evaluation value Dm calculated last and the AF evaluation value Dm-1 that is previous to the final AF evaluation value Dm (previous-to-final AF evaluation value Dm-1) are acquired.
- In a step ST11, it is determined whether or not the ratio of the final AF evaluation value Dm to the previous-to-final AF evaluation value Dm-1 is in the range of 0.99 to 1.01 inclusive. The purpose of this determination is to regard the position of the focus lens 211 as the focus position to thereby specify the focus position when the ratio of the final AF evaluation value Dm to the previous-to-final AF evaluation value Dm-1 is in the range of 100%±1%, because when the ratio is in this range, the focus lens 211 has been driven to a position corresponding to an AF evaluation value in a range near the peak Qk (
FIG. 14 ), where the slope of the AF evaluation value curve is gentle. - If it is determined in this step ST11 that the relationship 0.99≦Dm/Dm-1 ≦1.01 is satisfied, i.e., if the focus position of the focus lens 211 is specified based on the above-described history information of the AF evaluation values (the previous-to-final AF evaluation value Dm-1 and the final AF evaluation value Dm), the operation sequence proceeds to a step ST19. In contrast, if not so, the operation sequence proceeds to a step ST12.
- In the step ST12, it is determined whether or not the ratio of the final AF evaluation value Dm to the previous-to-final AF evaluation value Dm-1 is larger than one. If the relationship Dm/Dm-1>1 is satisfied, the operation sequence proceeds to a step ST13 based on a determination that the AF evaluation value is in a monotonic increase state. If the relationship Dm/Dm-1≦1 is satisfied, the operation sequence proceeds to the step ST19 based on a determination that the AF evaluation value is in a monotonic decrease state.
- In the step ST13, similarly to the step ST6, an AF evaluation value D1 is acquired based on outputs from the
non-divided G pixels 11 gb in the AF area in theimaging element 101. At the timing immediately after the completion of the focal adjustment by the phase difference AF, the above-described final AF evaluation value Dm is acquired as the AF evaluation value D1 from the history information of the AF evaluation value. - In a step ST14, additional driving of the focus lens 211 is carried out by 1Fδ equivalent to the focal depth (depth of field) in the same direction as the drive direction of the phase difference AF. Of the 1Fδ, F denotes the F number indicating the actual diaphragm regarding the imaging lens (imaging optical system) 2, and δ denotes the length twice the pixel pitch of the imaging element 101 (for example, when the pixel pitch is 6 μm, δ=12 μm).
- In a step ST15, similarly to the step ST6, an AF evaluation value D2 is acquired based on outputs from the
non-divided G pixels 11 gb in the AF area in theimaging element 101. - In a step ST 16, it is determined whether or not the ratio of the AF evaluation value D2 acquired in the step ST15 to the AF evaluation value D1 acquired in the step ST13 is in the range of 0.99 to 1.01 inclusive. The purpose of this determination is as follows. Specifically, when the AF evaluation value D2 is obtained through additional driving of the focus lens 211 by a drive amount W as shown in
FIG. 16 for example in the state in which the AF evaluation value has not reached the peak Qk (FIG. 14 ) but is monotonically increasing, if the ratio of the AF evaluation value D1 to the AF evaluation value D2 is in the range of 100%±1%, i.e., if the difference Ef between the AF evaluation values D2 and D1 is smaller than or equal to 1% of the AF evaluation value D1, the focus lens 211 has been driven to a position corresponding to an AF evaluation value in a gentle slope area near the peak Qk (FIG. 14 ). Therefore, through the determination in the step ST16, the position of the additionally-driven focus lens 211 is regarded as the focus position to thereby specify the focus position when the AF evaluation value ratio is in the range of 100%±1%. - If it is determined in the step ST16 that the relationship 0.99≦D2/D1≦1.01 is satisfied, the operation sequence proceeds to the step ST19. In contrast, if not so, the operation sequence proceeds to a step ST17 based on a determination that the focus lens 211 has not reached a position corresponding to an AF evaluation value near the peak Qk (
FIG. 14 ). - Through the operation of the steps ST13 to ST16, the following processing is executed. Specifically, if the focus position of the focus lens 211 is not specified based on the history information of the AF evaluation values acquired so far, additional driving of the focus lens 211 by a drive amount based on the focal depth relating to the
imaging lens 2 is carried out and an AF evaluation value is additionally acquired. Subsequently, the focus position of the focus lens 211 is specified based on the focal detection history information to which this additionally-acquired AF evaluation value (focal detection information) is added. - In the step ST17, it is determined whether or not the additional driving of the focus lens 211 in the step ST13 has been carried out n (e.g., n=3) times. The purpose of this determination is to stop the AF operation based on a determination that the focusing is difficult when the additional driving of the focus lens 211 is carried out several times. If the additional driving has been carried out n times, the operation sequence proceeds to the step ST19. If the number of times of additional driving is smaller than n, the operation sequence returns to the step ST13.
- In the step ST18, the focus lens 211 is back-driven to the position corresponding to the peak of the AF evaluation value. Specifically, if it is determined in the step ST12 that the ratio of the final AF evaluation value Dm to the previous-to-final AF evaluation value Dm-1 is lower than or equal to one, it is determined that the focus lens 211 has passed through the position corresponding to the peak of the AF evaluation value, and thus the focus position of the focus lens 211 is specified by using Equation (3). Furthermore, the focus lens 211, which has passed through the position (focus position) corresponding to the peak of the AF evaluation value through the phase difference AF, is back-driven to the specified focus position.
- That is, if the focus position of the focus lens 211 is specified based on the history information of the AF evaluation values acquired during the phase difference AF and the specified focus position is different from the focal detection position detected by the phase difference AF, the focus lens 211 is driven to this focus position through the operation of the step ST18.
- In the step ST19, the
imaging element 101 is set from the AF read mode to the live-view read mode. - Through the above-described operation of the
imaging device 1, focal detection processing by a phase difference detection system (phase difference AF) is executed based on charge signals obtained from the plural divided G pixels (second pixels) 11 gr. Furthermore, contrast AF different from the phase difference AF, live-view displaying, AE, and AWB are performed based on charge signals obtained from the plural non-divided pixels (first pixels) including thenon-divided G pixels 11 gb, theR pixels 11 r, and theB pixels 11 b. Therefore, specific processing necessary for camera functions other than the phase difference AF can be executed with high accuracy. - In addition, in the
imaging device 1, the dividedpixels 11 gr in which the pluralphotoelectric converters 111 to 118 are arranged have thephotoelectric converters - Furthermore, in the
imaging device 1, outputs from the divided pixels and the non-divided pixels are amplified by different gains in the AGC circuit of thesignal processor 52. Therefore, the output level of the divided pixels as well as that of the non-divided pixels can be set to the proper level. - Moreover, in the
imaging device 1, the respectivenon-divided pixels 11 gb have a color filter of the same color (green), which allows highly-accurate and proper phase difference AF. - In addition, in the
imaging device 1, the specific processing executed based on charge signal obtained from the non-divided pixels encompasses processing of contrast AF, processing relating to auto exposure control (AE), processing relating to auto white balance control (AWB), and processing of creating images relating to live-view displaying (preview displaying). Thus, the contrast AF, the AE control, and the AWB control can be carried out with high accuracy, and the live-view displaying can be performed properly. - In the above-described embodiment, it is not essential to carry out phase difference AF by use of an imaging element having divided pixels (divided G pixels), of which inside is divided. The phase difference AF may be carried out by using any of
imaging elements 101A and 101B shown inFIGS. 17 and 18 . -
FIG. 17 is a diagram for explaining the configuration of the imaging element 101A according to a modification example of the present invention. - On Gr lines (see the Gr lines L1 in
FIG. 5 ) of the imaging element 101A, a pair ofG pixels 11 g (11 gs, 11 gt) having agreen color filter 12 g are arranged to sandwich anR pixel 11 r having ared color filter 12 r. In theG pixel 11 gs, due to a slit SLa of a light-shieldingplate 13 a, a light beam Tb that has passed through a left-side part of an exit pupil Ep passes through thegreen color filter 12 g and forms an image on aphotoelectric converter 110. On the other hand, in theG pixel 11 gt, due to a slit SLb of a light-shieldingplate 13 b, a light beam Ta that has passed through a right-side part of the exit pupil Ep passes through thegreen color filter 12 g and forms an image on thephotoelectric converter 110. - If the light-reception data obtained from the
G pixels 11 gs and 11 gt with such a configuration are used as the above-described A-series data and B-series data, phase difference AF can be carried out similarly to the above-described dividedG pixels 11 gr. - Also in such an imaging element 101A, similarly to the
imaging element 101 of the above-described embodiment, phase difference AF is carried out by using the G pixels 11 (11 gs, 11 gt) on the Gr lines, while contrast AF, live-view displaying, AE, and AWB are carried out by using R pixels, B pixels, and normal G pixels that are arranged on Gb lines and in which the light-shieldingplates -
FIG. 18 is a diagram for explaining the configuration of theimaging element 101B according to another modification example. - On Gr lines (see the Gr lines L1 in
FIG. 5 ) of theimaging element 101B, a pair ofG pixels 11 g (11 gv, 11 gw) having agreen color filter 12 g are arranged to sandwich anR pixel 11 r having ared color filter 12 r. TheG pixel 11 gv is provided with a microlens ML of which top surface is provided with a light-shielding layer Qa obtained through black-coating with a pigment or paint for the entire surface other than a light transmissive area Pa equivalent to the slit SLa inFIG. 17 . Due to this microlens ML, in theG pixel 11 gv, a light beam Tb that has passed through a left-side part of an exit pupil Ep passes through thegreen color filter 12 g and forms an image on aphotoelectric converter 110. On the other hand, theG pixel 11 gw is provided with the microlens ML of which top surface is provided with a light-shielding layer Qb obtained through black-coating with a pigment or paint for the entire surface other than a light transmissive area Pb equivalent to the slit SLb inFIG. 17 . Due to this microlens ML, in theG pixel 11 gw, a light beam Ta that has passed through a right-side part of the exit pupil Ep passes through thegreen color filter 12 g and forms an image on thephotoelectric converter 110. - If the light-reception data obtained from the
G pixels 11 gv and 11 gw with such a configuration are used as the above-described A-series data and B-series data, phase difference AF can be carried out similarly to the above-described dividedG pixels 11 gr. - Also in such an
imaging element 101B, similarly to theimaging element 101 of the above-described embodiment, phase difference AF is carried out by using the G pixels 11 (11 gv, 11 gw) on the Gr lines, while contrast AF, live-view displaying, AE, and AWB are carried out by using R pixels, B pixels, and normal G pixels that are arranged on Gb lines and in which the light-shielding layers Qa and Qb are not formed on the microlens ML. This allows execution of specific processing necessary for camera functions other than the phase difference AF with high accuracy. - In the imaging device of the above-described embodiment, it is not essential that the
imaging lens 2 is freely detachable from thecamera body 10. Theimaging lens 2 may be fixed to thecamera body 10. - In the imaging element of the above-described embodiment, the inside of the G pixel does not necessarily need to be divided into eight areas as long as the pixel inside is divided into two or more areas. Furthermore, it is not essential to divide the G pixels. R pixels or B pixels may be divided.
- For the AF evaluation value of the above-described embodiment, it is not essential to calculate the total sum of the absolute values of the differences between adjacent
non-divided G pixels 11 gb. It is also possible to calculate the total sum of the absolute values of the squares of the differences between adjacent pixels. - In the operation of the imaging device of the above-described embodiment, it is not essential to determine in the step ST11 of
FIG. 15 whether or not the ratio of the final AF evaluation value Dm to the previous-to-final AF evaluation value Dm-1 is in the range of 100%±1%. For example, whether or not the ratio is in the range of 100%±3% may be determined. - In the operation of the imaging device of the above-described embodiment, it is not essential to carry out additional driving of the focus lens 211 by 1Fδ in the step ST14 of
FIG. 15 . Additional driving by 2Fδ may be carried out. That is, the drive amount is based on the focal depth relating to theimaging lens 2. - The divided G pixel in the above-described embodiment does not necessarily need to be divided into plural areas along the direction of the Gr lines L1 shown in
FIG. 5 (horizontal direction). It may be divided into plural areas along the vertical direction. In this case, phase difference AF is carried out based on the amount of the shift in the vertical direction regarding a pair of image sequences (A-series image sequence and B-series image sequence) obtained from the divided G pixels. - It should be noted that the present invention is not limited to the aforementioned embodiments and may be modified in various ways within the spirit of the invention.
Claims (8)
1. An imaging device having an imaging optical system, the device comprising:
(a) an imaging element configured to include a plurality of first pixels and a plurality of second pixels arranged along a predetermined direction;
(b) a first processor configured to execute focal detection processing by a phase difference detection system based on charge signals obtained from the plurality of second pixels; and
(c) a second processor configured to execute specific processing based on charge signals obtained from the plurality of first pixels, the specific processing being different from the focal detection processing by a phase difference detection system and being necessary for a function of the imaging device, wherein
the plurality of first pixels receive a subject light beam that has passed through an entire area of an exit pupil of the imaging optical system, and the plurality of second pixels receive subject light beams that have passed through a pair of partial areas of the exit pupil, and
the first processor creates a pair of image sequences based on charge signals from the second pixels that receive subject light beams that have passed through the pair of partial areas, and detects an amount of shift along the predetermined direction regarding the pair of image sequences, to thereby execute the focal detection processing by a phase difference detection system.
2. The imaging device according to claim 1 , wherein
in the second pixel, a plurality of photoelectric converters are arranged along the predetermined direction, and
the plurality of photoelectric converters include photoelectric converters that generate charge signals each relating to a respective one of the pair of image sequences.
3. The imaging device according to claim 1 , further comprising
(d) an amplifier configured to amplify a charge signal output from the imaging element, wherein
the amplifier amplifies a charge signal from the first pixel by a first amplification factor, and amplifies a charge signal from the second pixel by a second amplification factor different from the first amplification factor.
4. The imaging device according to claim 1 , wherein the plurality of second pixels each have a color filter of the same color.
5. The imaging device according to claim 1 , wherein the specific processing includes focal detection processing by a contrast detection system.
6. The imaging device according to claim 1 , further comprising:
(e) a display unit configured to display an image; and
(f) a display controller configured to cause the display unit to perform preview displaying of a subject based on an image signal produced by the imaging element before actual imaging, wherein
the specific processing includes processing of creating an image relating to the preview displaying.
7. The imaging device according to claim 1 , wherein the specific processing includes processing relating to auto exposure control.
8. The imaging device according to claim 1 , wherein the specific processing includes processing relating to auto white balance control.
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/244,598 US9143675B2 (en) | 2006-11-28 | 2014-04-03 | Imaging device having autofocus capability |
US14/847,400 US9986146B2 (en) | 2006-11-28 | 2015-09-08 | Imaging device having autofocus capability |
US15/967,876 US10375295B2 (en) | 2006-11-28 | 2018-05-01 | Imaging device having autofocus capability |
US16/503,770 US10674071B2 (en) | 2006-11-28 | 2019-07-05 | Imaging device having autofocus capability |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JPP2006-319783 | 2006-11-28 | ||
JP2006319783A JP4321579B2 (en) | 2006-11-28 | 2006-11-28 | Imaging device |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/244,598 Continuation US9143675B2 (en) | 2006-11-28 | 2014-04-03 | Imaging device having autofocus capability |
Publications (1)
Publication Number | Publication Date |
---|---|
US20080259202A1 true US20080259202A1 (en) | 2008-10-23 |
Family
ID=39487928
Family Applications (5)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/983,962 Abandoned US20080259202A1 (en) | 2006-11-28 | 2007-11-13 | Imaging device |
US14/244,598 Expired - Fee Related US9143675B2 (en) | 2006-11-28 | 2014-04-03 | Imaging device having autofocus capability |
US14/847,400 Expired - Fee Related US9986146B2 (en) | 2006-11-28 | 2015-09-08 | Imaging device having autofocus capability |
US15/967,876 Expired - Fee Related US10375295B2 (en) | 2006-11-28 | 2018-05-01 | Imaging device having autofocus capability |
US16/503,770 Active US10674071B2 (en) | 2006-11-28 | 2019-07-05 | Imaging device having autofocus capability |
Family Applications After (4)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/244,598 Expired - Fee Related US9143675B2 (en) | 2006-11-28 | 2014-04-03 | Imaging device having autofocus capability |
US14/847,400 Expired - Fee Related US9986146B2 (en) | 2006-11-28 | 2015-09-08 | Imaging device having autofocus capability |
US15/967,876 Expired - Fee Related US10375295B2 (en) | 2006-11-28 | 2018-05-01 | Imaging device having autofocus capability |
US16/503,770 Active US10674071B2 (en) | 2006-11-28 | 2019-07-05 | Imaging device having autofocus capability |
Country Status (4)
Country | Link |
---|---|
US (5) | US20080259202A1 (en) |
JP (1) | JP4321579B2 (en) |
CN (1) | CN101193209B (en) |
TW (1) | TWI375464B (en) |
Cited By (46)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090086083A1 (en) * | 2007-09-27 | 2009-04-02 | Keiji Kunishige | Imaging device, and control method for imaging device |
US20090110382A1 (en) * | 2007-10-26 | 2009-04-30 | Sony Corporation | Image capturing apparatus |
US20090110383A1 (en) * | 2007-10-29 | 2009-04-30 | Sony Corporation | Imaging device |
US20090148147A1 (en) * | 2007-12-10 | 2009-06-11 | Sony Corporation | Image-capturing apparatus |
US20090153705A1 (en) * | 2007-12-18 | 2009-06-18 | Sony Corporation | Image-capturing element and image-capturing apparatus |
US20100166405A1 (en) * | 2008-12-30 | 2010-07-01 | Samsung Electronics Co., Ltd. | Single lens reflex camera comprising a focus detecting apparatus and method of photographing |
US20100194967A1 (en) * | 2007-09-14 | 2010-08-05 | Canon Kabushiki Kaisha | Imaging apparatus |
US20100284679A1 (en) * | 2009-05-08 | 2010-11-11 | Sony Corporation | Imaging apparatus and imaging lens unit |
US20100309365A1 (en) * | 2009-06-05 | 2010-12-09 | Canon Kabushiki Kaisha | Image pickup apparatus having improved contrast autofocus accuracy |
US20110025904A1 (en) * | 2008-03-11 | 2011-02-03 | Canon Kabushiki Kaisha | Focus detection device and imaging apparatus having the same |
EP2430487A1 (en) * | 2009-05-12 | 2012-03-21 | Canon Kabushiki Kaisha | Focus detection apparatus |
EP2435879A1 (en) * | 2009-05-25 | 2012-04-04 | Canon Kabushiki Kaisha | Image capturing apparatus |
US20120133813A1 (en) * | 2010-11-29 | 2012-05-31 | Canon Kabushiki Kaisha | Image pickup apparatus |
US20120293706A1 (en) * | 2011-05-16 | 2012-11-22 | Samsung Electronics Co., Ltd. | Image pickup device, digital photographing apparatus using the image pickup device, auto-focusing method, and computer-readable medium for performing the auto-focusing method |
US20120300104A1 (en) * | 2011-05-26 | 2012-11-29 | Canon Kabushiki Kaisha | Image sensor and image capturing apparatus |
US20130038691A1 (en) * | 2011-08-12 | 2013-02-14 | Aptina Imaging Corporation | Asymmetric angular response pixels for single sensor stereo |
US20130278817A1 (en) * | 2012-04-19 | 2013-10-24 | Canon Kabushiki Kaisha | Ranging apparatus, ranging method and imaging system |
EP2257045A3 (en) * | 2009-05-26 | 2013-11-06 | Sony Corporation | Focus detection apparatus, image pickup device, and electronic camera |
US20140028895A1 (en) * | 2011-03-31 | 2014-01-30 | Fujifilm Corporation | Imaging apparatus and driving method thereof |
US20140049668A1 (en) * | 2011-04-28 | 2014-02-20 | Fujifilm Corporation | Imaging device and imaging method |
CN103733106A (en) * | 2011-07-25 | 2014-04-16 | 佳能株式会社 | Image pickup apparatus and control method thereof |
EP2330449A4 (en) * | 2008-09-24 | 2014-05-14 | Sony Corp | Imaging element and imaging device |
US8730380B2 (en) * | 2011-03-31 | 2014-05-20 | Fujifilm Corporation | Imaging device, method for controlling imaging device, and computer-readable storage medium |
US20140146218A1 (en) * | 2012-11-29 | 2014-05-29 | Canon Kabushiki Kaisha | Focus detection apparatus, image pickup apparatus, image pickup system, focus detection method, and non-transitory computer-readable storage medium |
US20150163395A1 (en) * | 2013-12-05 | 2015-06-11 | Canon Kabushiki Kaisha | Image capturing apparatus and control method thereof |
US20150163440A1 (en) * | 2013-12-10 | 2015-06-11 | Canon Kabushiki Kaisha | Image pickup apparatus that displays image based on signal output from image pickup device, method of controlling the same, and storage medium |
US20150185434A1 (en) * | 2011-07-25 | 2015-07-02 | Canon Kabushiki Kaisha | Image pickup apparatus, control method thereof, and program |
US20150195446A1 (en) * | 2014-01-07 | 2015-07-09 | Canon Kabushiki Kaisha | Imaging apparatus and its control method |
US20150215554A1 (en) * | 2014-01-29 | 2015-07-30 | Canon Kabushiki Kaisha | Solid-state imaging apparatus and imaging system using the same |
US20150253648A1 (en) * | 2012-11-22 | 2015-09-10 | Fujifilm Corporation | Imaging device, focusing method thereof, and non-transitory computer readable medium |
US9143675B2 (en) | 2006-11-28 | 2015-09-22 | Sony Corporation | Imaging device having autofocus capability |
US20150296128A1 (en) * | 2014-04-15 | 2015-10-15 | Canon Kabushiki Kaisha | Control apparatus and control method |
US9172862B2 (en) | 2011-10-31 | 2015-10-27 | Sony Corporation | Information processing device, information processing method, and program |
JP2015203774A (en) * | 2014-04-14 | 2015-11-16 | キヤノン株式会社 | Image-capturing device and control method thereof |
US9338380B2 (en) | 2014-06-30 | 2016-05-10 | Semiconductor Components Industries, Llc | Image processing methods for image sensors with phase detection pixels |
US9385148B2 (en) | 2012-01-13 | 2016-07-05 | Nikon Corporation | Solid-state imaging device and electronic camera |
US9432568B2 (en) * | 2014-06-30 | 2016-08-30 | Semiconductor Components Industries, Llc | Pixel arrangements for image sensors with phase detection pixels |
US9554115B2 (en) | 2012-02-27 | 2017-01-24 | Semiconductor Components Industries, Llc | Imaging pixels with depth sensing capabilities |
US9749556B2 (en) | 2015-03-24 | 2017-08-29 | Semiconductor Components Industries, Llc | Imaging systems having image sensor pixel arrays with phase detection capabilities |
US20170278890A1 (en) * | 2010-02-25 | 2017-09-28 | Nikon Corporation | Backside illumination image sensor and image-capturing device |
US9826183B2 (en) | 2012-03-30 | 2017-11-21 | Nikon Corporation | Image-capturing device and image sensor |
US9888198B2 (en) | 2014-06-03 | 2018-02-06 | Semiconductor Components Industries, Llc | Imaging systems having image sensor pixel arrays with sub-pixel resolution capabilities |
US9924094B2 (en) | 2013-11-25 | 2018-03-20 | Canon Kabushiki Kaisha | Image pickup apparatus capable of changing drive mode and image signal control method |
US20180146147A1 (en) * | 2016-11-18 | 2018-05-24 | Canon Kabushiki Kaisha | Image-capturing apparatus and control method therefor |
US10014336B2 (en) | 2011-01-28 | 2018-07-03 | Semiconductor Components Industries, Llc | Imagers with depth sensing capabilities |
US10917555B2 (en) | 2017-03-30 | 2021-02-09 | Sony Corporation | Imaging apparatus, focus control method, and focus determination method |
Families Citing this family (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5146295B2 (en) * | 2008-12-15 | 2013-02-20 | ソニー株式会社 | Imaging apparatus and focus control method |
JP6069817B2 (en) * | 2011-06-29 | 2017-02-01 | 株式会社ニコン | Focus adjustment device and imaging device provided with the same |
JP5979849B2 (en) * | 2011-11-21 | 2016-08-31 | キヤノン株式会社 | Imaging device and imaging apparatus |
JP5807125B2 (en) * | 2012-10-26 | 2015-11-10 | 富士フイルム株式会社 | Imaging apparatus and focus control method thereof |
JP6289017B2 (en) * | 2013-03-27 | 2018-03-07 | キヤノン株式会社 | Imaging apparatus, control method therefor, program, and storage medium |
JP6295526B2 (en) * | 2013-07-11 | 2018-03-20 | ソニー株式会社 | Solid-state imaging device and electronic apparatus |
CN105594197A (en) * | 2013-09-27 | 2016-05-18 | 富士胶片株式会社 | Imaging device and imaging method |
JP6223160B2 (en) * | 2013-12-10 | 2017-11-01 | キヤノン株式会社 | Imaging device, control method thereof, and control program |
JP6146293B2 (en) | 2013-12-25 | 2017-06-14 | ソニー株式会社 | Control device, control method, and control system |
JP6187244B2 (en) | 2013-12-25 | 2017-08-30 | ソニー株式会社 | Control apparatus, control method, and exposure control system |
CN104793910B (en) * | 2014-01-20 | 2018-11-09 | 联想(北京)有限公司 | A kind of method and electronic equipment of information processing |
JP6600170B2 (en) | 2014-07-07 | 2019-10-30 | キヤノン株式会社 | Image pickup device, control method therefor, and image pickup apparatus |
US9628695B2 (en) * | 2014-12-29 | 2017-04-18 | Intel Corporation | Method and system of lens shift correction for a camera array |
JP6672081B2 (en) | 2016-06-01 | 2020-03-25 | キヤノン株式会社 | Imaging device |
CN106101554B (en) * | 2016-07-29 | 2017-08-08 | 广东欧珀移动通信有限公司 | camera focusing method, device and terminal |
JP7071055B2 (en) * | 2017-02-24 | 2022-05-18 | キヤノン株式会社 | Image sensor and image sensor |
JP7227777B2 (en) | 2019-02-04 | 2023-02-22 | キヤノン株式会社 | Imaging device |
TWI772040B (en) * | 2021-05-27 | 2022-07-21 | 大陸商珠海凌煙閣芯片科技有限公司 | Object depth information acquistition method, device, computer device and storage media |
CN114125243A (en) * | 2021-11-30 | 2022-03-01 | 维沃移动通信有限公司 | Image sensor, camera module, electronic equipment and pixel information acquisition method |
Citations (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4783383A (en) * | 1986-12-02 | 1988-11-08 | Sanyo Electric Co., Ltd. | Sealed type battery provided with safety valve means and method of manufacturing same |
US5598964A (en) * | 1995-09-29 | 1997-02-04 | Motorola, Inc. | Apparatus for ultrasonic welding of battery assemblies |
US20010014215A1 (en) * | 2000-02-09 | 2001-08-16 | Olympus Optical Co., Ltd. | Distance measuring device |
US20020121652A1 (en) * | 2000-11-28 | 2002-09-05 | Ryo Yamasaki | Image pickup apparatus |
US6597868B2 (en) * | 2000-04-19 | 2003-07-22 | Canon Kabushiki Kaisha | Focus detecting device determining focusing state based on phase difference information |
US20040070366A1 (en) * | 2001-09-14 | 2004-04-15 | Toshio Takeshita | Battery pack |
US6750437B2 (en) * | 2000-08-28 | 2004-06-15 | Canon Kabushiki Kaisha | Image pickup apparatus that suitably adjusts a focus |
US6781632B1 (en) * | 1999-04-20 | 2004-08-24 | Olympus Corporation | Image pick-up apparatus capable of focus detection |
US6819360B1 (en) * | 1999-04-01 | 2004-11-16 | Olympus Corporation | Image pickup element and apparatus for focusing |
US20040240871A1 (en) * | 2003-03-14 | 2004-12-02 | Junichi Shinohara | Image inputting apparatus |
US6829008B1 (en) * | 1998-08-20 | 2004-12-07 | Canon Kabushiki Kaisha | Solid-state image sensing apparatus, control method therefor, image sensing apparatus, basic layout of photoelectric conversion cell, and storage medium |
US6897899B1 (en) * | 1999-05-10 | 2005-05-24 | Olympus Optical Co., Ltd. | Electronic image pickup apparatus |
US20070154200A1 (en) * | 2006-01-05 | 2007-07-05 | Nikon Corporation | Image sensor and image capturing device |
US20070206940A1 (en) * | 2006-03-01 | 2007-09-06 | Nikon Corporation | Focus adjustment device, imaging device and focus adjustment method |
US20090096886A1 (en) * | 2007-10-01 | 2009-04-16 | Nikon Corporation | Image-capturing device, camera, method for constructing image-capturing device and image-capturing method |
US20090128671A1 (en) * | 2007-11-16 | 2009-05-21 | Nikon Corporation | Imaging apparatus |
Family Cites Families (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS63259962A (en) | 1987-04-16 | 1988-10-27 | Sanyo Electric Co Ltd | Manufacture of enclosed type cell with safety valve |
JP3467047B2 (en) * | 1992-08-21 | 2003-11-17 | 富士写真フイルム株式会社 | Image processing apparatus and method, and video camera |
JP4007713B2 (en) | 1999-04-06 | 2007-11-14 | オリンパス株式会社 | Imaging device |
EP1071147A1 (en) | 1999-07-19 | 2001-01-24 | Toshiba Battery Co., Ltd. | Battery pack |
EP1175101B1 (en) | 2000-07-14 | 2013-11-13 | Texas Instruments Incorporated | Digital still camera system and method. |
JP2002189164A (en) * | 2000-12-21 | 2002-07-05 | Minolta Co Ltd | Optical system controller, optical system control method, and recording medium |
JP4972823B2 (en) | 2001-02-21 | 2012-07-11 | パナソニック株式会社 | Battery pack |
JP3591523B2 (en) | 2002-04-11 | 2004-11-24 | 日産自動車株式会社 | Battery pack |
JP4090276B2 (en) * | 2002-05-28 | 2008-05-28 | 富士フイルム株式会社 | Digital camera |
JP3788393B2 (en) | 2002-06-10 | 2006-06-21 | ソニー株式会社 | Digital still camera device, video camera device and information terminal device |
JP2005092085A (en) | 2003-09-19 | 2005-04-07 | Canon Inc | Focus detecting method and focusing method, and focus detecting device and focusing device |
US7773146B2 (en) * | 2004-06-15 | 2010-08-10 | Canon Kabushiki Kaisha | Focus control apparatus and optical apparatus |
TWI300159B (en) | 2004-12-24 | 2008-08-21 | Sony Taiwan Ltd | Camera system |
EP1684503B1 (en) * | 2005-01-25 | 2016-01-13 | Canon Kabushiki Kaisha | Camera and autofocus control method therefor |
US7620312B2 (en) | 2005-04-11 | 2009-11-17 | Canon Kabushiki Kaisha | Focus detection apparatus and signal processing method for focus detection |
JP4321579B2 (en) | 2006-11-28 | 2009-08-26 | ソニー株式会社 | Imaging device |
JP2009069255A (en) * | 2007-09-11 | 2009-04-02 | Sony Corp | Imaging device and focusing control method |
-
2006
- 2006-11-28 JP JP2006319783A patent/JP4321579B2/en not_active Expired - Fee Related
-
2007
- 2007-10-05 TW TW096137557A patent/TWI375464B/en not_active IP Right Cessation
- 2007-11-13 US US11/983,962 patent/US20080259202A1/en not_active Abandoned
- 2007-11-28 CN CN2007101965219A patent/CN101193209B/en not_active Expired - Fee Related
-
2014
- 2014-04-03 US US14/244,598 patent/US9143675B2/en not_active Expired - Fee Related
-
2015
- 2015-09-08 US US14/847,400 patent/US9986146B2/en not_active Expired - Fee Related
-
2018
- 2018-05-01 US US15/967,876 patent/US10375295B2/en not_active Expired - Fee Related
-
2019
- 2019-07-05 US US16/503,770 patent/US10674071B2/en active Active
Patent Citations (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4783383A (en) * | 1986-12-02 | 1988-11-08 | Sanyo Electric Co., Ltd. | Sealed type battery provided with safety valve means and method of manufacturing same |
US5598964A (en) * | 1995-09-29 | 1997-02-04 | Motorola, Inc. | Apparatus for ultrasonic welding of battery assemblies |
US6829008B1 (en) * | 1998-08-20 | 2004-12-07 | Canon Kabushiki Kaisha | Solid-state image sensing apparatus, control method therefor, image sensing apparatus, basic layout of photoelectric conversion cell, and storage medium |
US6819360B1 (en) * | 1999-04-01 | 2004-11-16 | Olympus Corporation | Image pickup element and apparatus for focusing |
US6781632B1 (en) * | 1999-04-20 | 2004-08-24 | Olympus Corporation | Image pick-up apparatus capable of focus detection |
US6897899B1 (en) * | 1999-05-10 | 2005-05-24 | Olympus Optical Co., Ltd. | Electronic image pickup apparatus |
US20010014215A1 (en) * | 2000-02-09 | 2001-08-16 | Olympus Optical Co., Ltd. | Distance measuring device |
US6597868B2 (en) * | 2000-04-19 | 2003-07-22 | Canon Kabushiki Kaisha | Focus detecting device determining focusing state based on phase difference information |
US6750437B2 (en) * | 2000-08-28 | 2004-06-15 | Canon Kabushiki Kaisha | Image pickup apparatus that suitably adjusts a focus |
US20020121652A1 (en) * | 2000-11-28 | 2002-09-05 | Ryo Yamasaki | Image pickup apparatus |
US20040070366A1 (en) * | 2001-09-14 | 2004-04-15 | Toshio Takeshita | Battery pack |
US20040240871A1 (en) * | 2003-03-14 | 2004-12-02 | Junichi Shinohara | Image inputting apparatus |
US20070154200A1 (en) * | 2006-01-05 | 2007-07-05 | Nikon Corporation | Image sensor and image capturing device |
US20070206940A1 (en) * | 2006-03-01 | 2007-09-06 | Nikon Corporation | Focus adjustment device, imaging device and focus adjustment method |
US20090096886A1 (en) * | 2007-10-01 | 2009-04-16 | Nikon Corporation | Image-capturing device, camera, method for constructing image-capturing device and image-capturing method |
US20090128671A1 (en) * | 2007-11-16 | 2009-05-21 | Nikon Corporation | Imaging apparatus |
Cited By (102)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10674071B2 (en) | 2006-11-28 | 2020-06-02 | Sony Corporation | Imaging device having autofocus capability |
US10375295B2 (en) | 2006-11-28 | 2019-08-06 | Sony Corporation | Imaging device having autofocus capability |
US9143675B2 (en) | 2006-11-28 | 2015-09-22 | Sony Corporation | Imaging device having autofocus capability |
US20100194967A1 (en) * | 2007-09-14 | 2010-08-05 | Canon Kabushiki Kaisha | Imaging apparatus |
US8212917B2 (en) * | 2007-09-14 | 2012-07-03 | Canon Kabushiki Kaisha | Imaging apparatus |
US8405759B2 (en) * | 2007-09-27 | 2013-03-26 | Olympus Imagining Corp. | Imaging device with contrast AF, and control method for imaging device with contrast AF |
US20090086083A1 (en) * | 2007-09-27 | 2009-04-02 | Keiji Kunishige | Imaging device, and control method for imaging device |
US20090110382A1 (en) * | 2007-10-26 | 2009-04-30 | Sony Corporation | Image capturing apparatus |
US7844173B2 (en) | 2007-10-26 | 2010-11-30 | Sony Corporation | Image capturing apparatus |
US7920782B2 (en) | 2007-10-29 | 2011-04-05 | Sony Corporation | Imaging device |
US20090110383A1 (en) * | 2007-10-29 | 2009-04-30 | Sony Corporation | Imaging device |
US8218962B2 (en) * | 2007-12-10 | 2012-07-10 | Sony Corporation | Image-capturing apparatus |
US20090148147A1 (en) * | 2007-12-10 | 2009-06-11 | Sony Corporation | Image-capturing apparatus |
US20090153705A1 (en) * | 2007-12-18 | 2009-06-18 | Sony Corporation | Image-capturing element and image-capturing apparatus |
US20110025904A1 (en) * | 2008-03-11 | 2011-02-03 | Canon Kabushiki Kaisha | Focus detection device and imaging apparatus having the same |
US8711270B2 (en) * | 2008-03-11 | 2014-04-29 | Canon Kabushiki Kaisha | Focus detection device and imaging apparatus having the same |
EP2330449A4 (en) * | 2008-09-24 | 2014-05-14 | Sony Corp | Imaging element and imaging device |
US8059954B2 (en) * | 2008-12-30 | 2011-11-15 | Samsung Electronics Co., Ltd. | Single lens reflex camera comprising a focus detecting apparatus and method of photographing |
US20100166405A1 (en) * | 2008-12-30 | 2010-07-01 | Samsung Electronics Co., Ltd. | Single lens reflex camera comprising a focus detecting apparatus and method of photographing |
US8064762B2 (en) * | 2009-05-08 | 2011-11-22 | Sony Corporation | Imaging apparatus and imaging lens unit |
US20100284679A1 (en) * | 2009-05-08 | 2010-11-11 | Sony Corporation | Imaging apparatus and imaging lens unit |
EP2430487A4 (en) * | 2009-05-12 | 2013-06-12 | Canon Kk | Focus detection apparatus |
EP2430487A1 (en) * | 2009-05-12 | 2012-03-21 | Canon Kabushiki Kaisha | Focus detection apparatus |
US8704933B2 (en) | 2009-05-12 | 2014-04-22 | Canon Kabushiki Kaisha | Focus detection apparatus |
EP2435879A1 (en) * | 2009-05-25 | 2012-04-04 | Canon Kabushiki Kaisha | Image capturing apparatus |
EP2435879A4 (en) * | 2009-05-25 | 2013-04-03 | Canon Kk | Image capturing apparatus |
US8890994B2 (en) | 2009-05-25 | 2014-11-18 | Canon Kabushiki Kaisha | Image capturing apparatus |
EP2257045A3 (en) * | 2009-05-26 | 2013-11-06 | Sony Corporation | Focus detection apparatus, image pickup device, and electronic camera |
US8264592B2 (en) * | 2009-06-05 | 2012-09-11 | Canon Kabushiki Kaisha | Image pickup apparatus having improved contrast autofocus accuracy |
US8773574B2 (en) * | 2009-06-05 | 2014-07-08 | Canon Kabushiki Kaisha | Image pickup apparatus having improved contrast autofocus accuracy |
US20100309365A1 (en) * | 2009-06-05 | 2010-12-09 | Canon Kabushiki Kaisha | Image pickup apparatus having improved contrast autofocus accuracy |
US10777595B2 (en) * | 2010-02-25 | 2020-09-15 | Nikon Corporation | Backside illumination image sensor and image-capturing device |
US20190140009A1 (en) * | 2010-02-25 | 2019-05-09 | Nikon Corporation | Backside illumination image sensor and image-capturing device |
US10204949B2 (en) * | 2010-02-25 | 2019-02-12 | Nikon Corporation | Backside illumination image sensor and image-capturing device |
US11601624B2 (en) * | 2010-02-25 | 2023-03-07 | Nikon Corporation | Backside illumination image sensor and image-capturing device |
US11979672B2 (en) | 2010-02-25 | 2024-05-07 | Nikon Corporation | Backside illumination image sensor and image-capturing device |
US20170278890A1 (en) * | 2010-02-25 | 2017-09-28 | Nikon Corporation | Backside illumination image sensor and image-capturing device |
US20120133813A1 (en) * | 2010-11-29 | 2012-05-31 | Canon Kabushiki Kaisha | Image pickup apparatus |
US10014336B2 (en) | 2011-01-28 | 2018-07-03 | Semiconductor Components Industries, Llc | Imagers with depth sensing capabilities |
US9106824B2 (en) * | 2011-03-31 | 2015-08-11 | Fujifilm Corporation | Imaging apparatus and driving method selecting one of a phase difference AF mode and a contrast AF mode |
US20140028895A1 (en) * | 2011-03-31 | 2014-01-30 | Fujifilm Corporation | Imaging apparatus and driving method thereof |
US8730380B2 (en) * | 2011-03-31 | 2014-05-20 | Fujifilm Corporation | Imaging device, method for controlling imaging device, and computer-readable storage medium |
US8830384B2 (en) * | 2011-04-28 | 2014-09-09 | Fujifilm Corporation | Imaging device and imaging method |
US20140049668A1 (en) * | 2011-04-28 | 2014-02-20 | Fujifilm Corporation | Imaging device and imaging method |
US20120293706A1 (en) * | 2011-05-16 | 2012-11-22 | Samsung Electronics Co., Ltd. | Image pickup device, digital photographing apparatus using the image pickup device, auto-focusing method, and computer-readable medium for performing the auto-focusing method |
US9215389B2 (en) * | 2011-05-16 | 2015-12-15 | Samsung Electronics Co., Ltd. | Image pickup device, digital photographing apparatus using the image pickup device, auto-focusing method, and computer-readable medium for performing the auto-focusing method |
US20150172577A1 (en) * | 2011-05-26 | 2015-06-18 | Canon Kabushiki Kaisha | Image sensor and image capturing apparatus |
US9204067B2 (en) * | 2011-05-26 | 2015-12-01 | Canon Kabushiki Kaisha | Image sensor and image capturing apparatus |
US20120300104A1 (en) * | 2011-05-26 | 2012-11-29 | Canon Kabushiki Kaisha | Image sensor and image capturing apparatus |
US9001262B2 (en) * | 2011-05-26 | 2015-04-07 | Canon Kabushiki Kaisha | Image sensor and image capturing apparatus |
CN103733106A (en) * | 2011-07-25 | 2014-04-16 | 佳能株式会社 | Image pickup apparatus and control method thereof |
DE112012003106B4 (en) | 2011-07-25 | 2019-07-18 | Canon Kabushiki Kaisha | Image pickup device and control method for this |
GB2506324B (en) * | 2011-07-25 | 2018-04-18 | Canon Kk | Image pickup apparatus and control method thereof |
US20140192220A1 (en) * | 2011-07-25 | 2014-07-10 | Canon Kabushiki Kaisha | Image pickup apparatus and control method thereof |
US20150185434A1 (en) * | 2011-07-25 | 2015-07-02 | Canon Kabushiki Kaisha | Image pickup apparatus, control method thereof, and program |
US9279955B2 (en) * | 2011-07-25 | 2016-03-08 | Canon Kabushiki Kaisha | Image pickup apparatus, control method thereof, and program |
US10009533B2 (en) * | 2011-07-25 | 2018-06-26 | Canon Kabushiki Kaisha | Image pickup apparatus and control method thereof with contrast focus evaluation |
US20130038691A1 (en) * | 2011-08-12 | 2013-02-14 | Aptina Imaging Corporation | Asymmetric angular response pixels for single sensor stereo |
US20180288398A1 (en) * | 2011-08-12 | 2018-10-04 | Semiconductor Components Industries, Llc | Asymmetric angular response pixels for singl sensor stereo |
US10015471B2 (en) * | 2011-08-12 | 2018-07-03 | Semiconductor Components Industries, Llc | Asymmetric angular response pixels for single sensor stereo |
US9172862B2 (en) | 2011-10-31 | 2015-10-27 | Sony Corporation | Information processing device, information processing method, and program |
US10674102B2 (en) | 2012-01-13 | 2020-06-02 | Nikon Corporation | Solid-state imaging device and electronic camera |
US9385148B2 (en) | 2012-01-13 | 2016-07-05 | Nikon Corporation | Solid-state imaging device and electronic camera |
US9654709B2 (en) | 2012-01-13 | 2017-05-16 | Nikon Corporation | Solid-state imaging device and electronic camera |
US11588991B2 (en) | 2012-01-13 | 2023-02-21 | Nikon Corporation | Solid-state imaging device and electronic camera |
US10158843B2 (en) * | 2012-02-27 | 2018-12-18 | Semiconductor Components Industries, Llc | Imaging pixels with depth sensing capabilities |
US9554115B2 (en) | 2012-02-27 | 2017-01-24 | Semiconductor Components Industries, Llc | Imaging pixels with depth sensing capabilities |
US20170094260A1 (en) * | 2012-02-27 | 2017-03-30 | Semiconductor Components Industries, Llc | Imaging pixels with depth sensing capabilities |
US20190089944A1 (en) * | 2012-02-27 | 2019-03-21 | Semiconductor Components Industries, Llc | Imaging pixels with depth sensing capabilities |
US10389959B2 (en) | 2012-03-30 | 2019-08-20 | Nikon Corporation | Image-capturing device and image sensor |
US9826183B2 (en) | 2012-03-30 | 2017-11-21 | Nikon Corporation | Image-capturing device and image sensor |
US20130278817A1 (en) * | 2012-04-19 | 2013-10-24 | Canon Kabushiki Kaisha | Ranging apparatus, ranging method and imaging system |
US8817168B2 (en) * | 2012-04-19 | 2014-08-26 | Canon Kabushiki Kaisha | Ranging apparatus, ranging method and imaging system |
US20150253648A1 (en) * | 2012-11-22 | 2015-09-10 | Fujifilm Corporation | Imaging device, focusing method thereof, and non-transitory computer readable medium |
US9547217B2 (en) * | 2012-11-22 | 2017-01-17 | Fujifilm Corporation | Device and, method for controlling a focusing operation |
US9380202B2 (en) | 2012-11-29 | 2016-06-28 | Canon Kabushiki Kaisha | Focus detection apparatus, image pickup apparatus, image pickup system, focus detection method, and non-transitory computer-readable storage medium |
US8988595B2 (en) * | 2012-11-29 | 2015-03-24 | Canon Kabushiki Kaisha | Focus detection apparatus, image pickup apparatus, image pickup system, focus detection method, and non-transitory computer-readable storage medium |
US20140146218A1 (en) * | 2012-11-29 | 2014-05-29 | Canon Kabushiki Kaisha | Focus detection apparatus, image pickup apparatus, image pickup system, focus detection method, and non-transitory computer-readable storage medium |
GB2534040B (en) * | 2013-11-25 | 2018-07-04 | Canon Kk | Image pickup apparatus capable of changing drive mode and image signal control method |
US9924094B2 (en) | 2013-11-25 | 2018-03-20 | Canon Kabushiki Kaisha | Image pickup apparatus capable of changing drive mode and image signal control method |
US9826140B2 (en) * | 2013-12-05 | 2017-11-21 | Canon Kabushiki Kaisha | Image capturing apparatus and control method thereof |
US20150163395A1 (en) * | 2013-12-05 | 2015-06-11 | Canon Kabushiki Kaisha | Image capturing apparatus and control method thereof |
US20150163440A1 (en) * | 2013-12-10 | 2015-06-11 | Canon Kabushiki Kaisha | Image pickup apparatus that displays image based on signal output from image pickup device, method of controlling the same, and storage medium |
US20180077398A1 (en) * | 2013-12-10 | 2018-03-15 | Canon Kabushiki Kaisha | Image pickup apparatus that displays image based on signal output from image pickup device, method of controlling the same, and storage medium |
US10349028B2 (en) * | 2013-12-10 | 2019-07-09 | Canon Kabushiki Kaisha | Image pickup apparatus that displays image based on signal output from image pickup device, method of controlling the same, and storage medium |
US9854216B2 (en) * | 2013-12-10 | 2017-12-26 | Canon Kabushiki Kaisha | Image pickup apparatus that displays image based on signal output from image pickup device, method of controlling the same, and storage medium |
US9363429B2 (en) * | 2014-01-07 | 2016-06-07 | Canon Kabushiki Kaisha | Imaging apparatus and its control method |
US20150195446A1 (en) * | 2014-01-07 | 2015-07-09 | Canon Kabushiki Kaisha | Imaging apparatus and its control method |
US20160316135A1 (en) * | 2014-01-07 | 2016-10-27 | Canon Kabushiki Kaisha | Imaging apparatus and its control method |
US9621789B2 (en) * | 2014-01-07 | 2017-04-11 | Canon Kabushiki Kaisha | Imaging apparatus and its control method |
US20150215554A1 (en) * | 2014-01-29 | 2015-07-30 | Canon Kabushiki Kaisha | Solid-state imaging apparatus and imaging system using the same |
US9544493B2 (en) * | 2014-01-29 | 2017-01-10 | Canon Kabushiki Kaisha | Solid-state imaging apparatus and imaging system using the same |
JP2015203774A (en) * | 2014-04-14 | 2015-11-16 | キヤノン株式会社 | Image-capturing device and control method thereof |
US20150296128A1 (en) * | 2014-04-15 | 2015-10-15 | Canon Kabushiki Kaisha | Control apparatus and control method |
US9300862B2 (en) * | 2014-04-15 | 2016-03-29 | Canon Kabushiki Kaisha | Control apparatus and control method |
US9888198B2 (en) | 2014-06-03 | 2018-02-06 | Semiconductor Components Industries, Llc | Imaging systems having image sensor pixel arrays with sub-pixel resolution capabilities |
US9338380B2 (en) | 2014-06-30 | 2016-05-10 | Semiconductor Components Industries, Llc | Image processing methods for image sensors with phase detection pixels |
US9432568B2 (en) * | 2014-06-30 | 2016-08-30 | Semiconductor Components Industries, Llc | Pixel arrangements for image sensors with phase detection pixels |
US9749556B2 (en) | 2015-03-24 | 2017-08-29 | Semiconductor Components Industries, Llc | Imaging systems having image sensor pixel arrays with phase detection capabilities |
US10623671B2 (en) * | 2016-11-18 | 2020-04-14 | Canon Kabushiki Kaisha | Image-capturing apparatus and control method therefor |
US20180146147A1 (en) * | 2016-11-18 | 2018-05-24 | Canon Kabushiki Kaisha | Image-capturing apparatus and control method therefor |
US10917555B2 (en) | 2017-03-30 | 2021-02-09 | Sony Corporation | Imaging apparatus, focus control method, and focus determination method |
Also Published As
Publication number | Publication date |
---|---|
CN101193209B (en) | 2012-09-05 |
US20150381881A1 (en) | 2015-12-31 |
TWI375464B (en) | 2012-10-21 |
JP4321579B2 (en) | 2009-08-26 |
TW200824444A (en) | 2008-06-01 |
US20180249067A1 (en) | 2018-08-30 |
CN101193209A (en) | 2008-06-04 |
US9143675B2 (en) | 2015-09-22 |
US20140218592A1 (en) | 2014-08-07 |
US10674071B2 (en) | 2020-06-02 |
JP2008134389A (en) | 2008-06-12 |
US20190335090A1 (en) | 2019-10-31 |
US10375295B2 (en) | 2019-08-06 |
US9986146B2 (en) | 2018-05-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10674071B2 (en) | Imaging device having autofocus capability | |
EP2249192B1 (en) | Imaging device, and imaging element | |
US7822334B2 (en) | Imaging device and in-focus control method | |
JP5146295B2 (en) | Imaging apparatus and focus control method | |
JP4285546B2 (en) | Imaging system, imaging apparatus, and interchangeable lens | |
US8218962B2 (en) | Image-capturing apparatus | |
US7920782B2 (en) | Imaging device | |
JP5003132B2 (en) | Imaging device and imaging apparatus | |
US8582019B2 (en) | Image pickup element and image pickup device | |
JP5109641B2 (en) | Imaging device and imaging apparatus | |
US7999217B2 (en) | Image-capturing element including photoelectric conversion cells that pupil-divide object light and output a ranging signal | |
JP2008134390A (en) | Imaging apparatus | |
JP2008134413A (en) | Imaging apparatus | |
JP5157525B2 (en) | Imaging device | |
JP2008039976A (en) | Imaging apparatus | |
JP2009272735A (en) | Imaging apparatus, photometric value correction method and photometric value correction program | |
JP2009150978A (en) | Imaging sensor and imaging device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SONY CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:FUJII, SHINICHI;REEL/FRAME:020489/0870 Effective date: 20070920 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |