JP6239857B2 - Imaging apparatus and control method thereof - Google Patents

Imaging apparatus and control method thereof Download PDF

Info

Publication number
JP6239857B2
JP6239857B2 JP2013101713A JP2013101713A JP6239857B2 JP 6239857 B2 JP6239857 B2 JP 6239857B2 JP 2013101713 A JP2013101713 A JP 2013101713A JP 2013101713 A JP2013101713 A JP 2013101713A JP 6239857 B2 JP6239857 B2 JP 6239857B2
Authority
JP
Japan
Prior art keywords
focus detection
imaging
pixel
focus
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
JP2013101713A
Other languages
Japanese (ja)
Other versions
JP2014222291A (en
Inventor
福田 浩一
浩一 福田
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Canon Inc
Original Assignee
Canon Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Canon Inc filed Critical Canon Inc
Priority to JP2013101713A priority Critical patent/JP6239857B2/en
Publication of JP2014222291A publication Critical patent/JP2014222291A/en
Application granted granted Critical
Publication of JP6239857B2 publication Critical patent/JP6239857B2/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Focusing (AREA)
  • Automatic Focus Adjustment (AREA)
  • Studio Devices (AREA)

Description

本発明は、撮像装置およびその制御方法に関する。   The present invention relates to an imaging apparatus and a control method thereof.

撮像装置の焦点検出方法の1つに、撮像素子に形成された焦点検出画素により位相差検出方式の焦点検出を行う撮像面位相差検出方式がある。   One of the focus detection methods of the imaging apparatus is an imaging plane phase difference detection method in which focus detection of a phase difference detection method is performed by focus detection pixels formed on an image sensor.

特許文献1には、1つの画素に、1つのマイクロレンズと複数に分割された光電変換部が形成された2次元撮像素子を用いた撮像装置が開示されている。1つのマイクロレンズを介して、撮影レンズの射出瞳の異なる領域を受光するように光電変換部を分割する。そして、焦点検出領域内に含まれる複数の画素の、同じ分割位置の光電変換部で得られた複数の信号から1つの像信号を生成することで、1対の像信号を生成する。この1対の像信号の像ずれ量から、位相差検出方式の焦点検出を行うことができる。また、特許文献2では、1つの画素に設けられた複数の光電変換部で得られる信号を加算することにより、その画素における撮像信号を生成することが開示されている。   Patent Document 1 discloses an imaging apparatus using a two-dimensional imaging element in which one microlens and a plurality of divided photoelectric conversion units are formed in one pixel. The photoelectric conversion unit is divided so as to receive different regions of the exit pupil of the photographing lens through one microlens. A pair of image signals is generated by generating one image signal from a plurality of signals obtained by the photoelectric conversion units at the same division position of a plurality of pixels included in the focus detection region. From the image shift amount of the pair of image signals, focus detection by the phase difference detection method can be performed. Japanese Patent Application Laid-Open No. 2004-228561 discloses generating an imaging signal in a pixel by adding signals obtained by a plurality of photoelectric conversion units provided in one pixel.

特許文献3には、複数の画素からなる2次元撮像素子の一部の画素を複数対の焦点検出画素とした撮像装置が開示されている。1対の焦点検出画素は開口部を有する遮光層の配置により、各焦点検出画素が撮影レンズの射出瞳の異なる領域を受光するように構成されている。2次元撮像素子の大部分に配置された撮像画素で撮像信号を取得する一方で、一部に配置された焦点検出画素から1対の像信号(焦点検出用信号)から像ずれ量を求めて、位相差検出方式の焦点検出を行うことが開示されている。   Patent Document 3 discloses an imaging apparatus in which some pixels of a two-dimensional imaging element composed of a plurality of pixels are used as a plurality of pairs of focus detection pixels. The pair of focus detection pixels are configured such that each focus detection pixel receives a different region of the exit pupil of the photographing lens by arranging a light shielding layer having an opening. While acquiring an imaging signal with imaging pixels arranged in most of the two-dimensional imaging device, an image shift amount is obtained from a pair of image signals (focus detection signals) from the focus detection pixels arranged in a part. It is disclosed that focus detection is performed using a phase difference detection method.

撮像面位相差検出方式の焦点検出においては、撮像素子に形成された焦点検出画素によりデフォーカス方向とデフォーカス量を同時に検出することが可能であり、高速に焦点調節を行うことができる。また、従来、位相差検出方式の焦点検出を行うために用いられていたAFセンサを用いる必要がない。   In focus detection using the imaging surface phase difference detection method, it is possible to detect the defocus direction and the defocus amount at the same time by focus detection pixels formed on the image sensor, and focus adjustment can be performed at high speed. Further, it is not necessary to use an AF sensor that has been conventionally used to perform focus detection by a phase difference detection method.

米国特許第4410804号U.S. Pat.No. 4,410,804 特開2001−083407号公報JP 2001-083407 A 特開2000−156823号公報JP 2000-156823 A

しかしながら、撮像面位相差検出方式では、焦点検出信号(像信号)を生成するために用いる焦点検出画素が受光する光束と、撮像信号を生成するために用いる撮像画素が受光する光束とが異なる。そのため、撮影レンズの各収差(球面収差、非点収差、コマ収差など)の焦点検出信号への影響と撮像信号への影響が異なる。そのため、焦点検出信号から算出される合焦位置と、撮像信号を得るために最良な合焦位置とが一致しないという課題がある。   However, in the imaging surface phase difference detection method, a light beam received by a focus detection pixel used for generating a focus detection signal (image signal) is different from a light beam received by an imaging pixel used for generating an imaging signal. Therefore, the influence of each aberration (spherical aberration, astigmatism, coma aberration, etc.) of the photographing lens on the focus detection signal is different from the influence on the imaging signal. For this reason, there is a problem that the in-focus position calculated from the focus detection signal does not match the best in-focus position for obtaining the imaging signal.

本発明は、このような従来技術の課題に鑑みてなされたものである。本発明は、焦点検出画素から得られる信号に基づいて位相差検出方式による自動焦点検出を行う撮像装置およびその制御方法において、撮影レンズの収差が焦点検出結果に与える影響を軽減することを目的とする。   The present invention has been made in view of such a problem of the prior art. An object of the present invention is to reduce the influence of aberration of a photographing lens on a focus detection result in an imaging apparatus that performs automatic focus detection by a phase difference detection method based on a signal obtained from a focus detection pixel. To do.

上述の目的は、結像光学系の射出瞳の第1部分を通過する光束を受光する複数の第1焦点検出用画素と、第1部分と異なる第2部分を通過する光束を受光する複数の第2焦点検出用画素とを有する撮像素子と、撮像素子の焦点検出領域に関連付けられた領域の第1焦点検出用画素から得られる信号から第1焦点検出信号を、領域の第2焦点検出用画素から得られる信号から第2焦点検出信号をそれぞれ生成する第1生成手段と、第1焦点検出信号と第2焦点検出信号との像ずれ量から、結像光学系のデフォーカス量を取得する第1取得手段と、第1焦点検出用画素から得られる信号と、第2焦点検出用画素から得られる信号とを用いて、記録媒体に記録する記録画像を生成する第2生成手段と、結像光学系の状態に関する情報と撮像素子に関する情報との組み合わせに応じた補正係数と、焦点検出領域の像高とから、デフォーカス量の補正値であって、第1取得手段にて取得されたデフォーカス量に基づく合焦位置と、記録画像の合焦位置との差を補正するための補正値を取得する第2取得手段と、補正値によってデフォーカス量を補正し、結像光学系の焦点調節に用いるデフォーカス量を取得する補正手段と、を有することを特徴とする撮像装置によって達成される。 The above-described object is to provide a plurality of first focus detection pixels that receive a light beam that passes through the first part of the exit pupil of the imaging optical system, and a plurality of light beams that pass through a second part different from the first part. A first focus detection signal is obtained from a signal obtained from an image sensor having a second focus detection pixel and a first focus detection pixel in a region associated with the focus detection region of the image sensor. The defocus amount of the imaging optical system is acquired from the first generation means for generating the second focus detection signal from the signal obtained from the pixel, and the image shift amount between the first focus detection signal and the second focus detection signal. A first acquisition unit, a second generation unit configured to generate a recording image to be recorded on a recording medium, using a signal obtained from the first focus detection pixel and a signal obtained from the second focus detection pixel; Information on the state of the image optical system and the image sensor A correction coefficient corresponding to a combination of that information, the image height of the focus detection area, a correction value of the defocus amount, the focusing position based on the defocus amount obtained by the first obtaining means, Second acquisition means for acquiring a correction value for correcting a difference from the in-focus position of the recorded image, and correcting the defocus amount by the correction value to acquire the defocus amount used for focus adjustment of the imaging optical system. And an correcting device.

このような構成により、本発明によれば、焦点検出画素から得られる信号に基づいて位相差検出方式による自動焦点検出を行う撮像装置およびその制御方法において、撮影レンズの収差が焦点検出結果に与える影響を軽減することができる。   With such a configuration, according to the present invention, in the imaging device that performs automatic focus detection by the phase difference detection method based on the signal obtained from the focus detection pixel and the control method thereof, the aberration of the photographic lens gives the focus detection result. The impact can be reduced.

本発明の実施形態に係る撮像装置の一例としてのデジタルスチルカメラの機能構成例を示す図1 is a diagram illustrating an example of a functional configuration of a digital still camera as an example of an imaging apparatus according to an embodiment of the present invention. 図1の撮像素子107における撮像画素と焦点検出画素の配置例を模式的に示す図The figure which shows typically the example of arrangement | positioning of the imaging pixel and focus detection pixel in the image pick-up element 107 of FIG. 図2に示した撮像画素の構成を模式的に示した平面図および断面図FIG. 2 is a plan view and a cross-sectional view schematically showing the configuration of the image pickup pixel shown in FIG. 図3に示した画素構造と瞳分割との対応関係を示した概略説明図Schematic explanatory diagram showing the correspondence between the pixel structure shown in FIG. 3 and pupil division 図4に示した結像光学系の射出瞳面のX軸に沿った焦点検出画素の瞳強度分布の例を示す図The figure which shows the example of the pupil intensity distribution of the focus detection pixel along the X-axis of the exit pupil plane of the imaging optical system shown in FIG. (a)は実施形態の撮像素子と瞳分割との対応関係を示した図、(b)は実施形態におけるデフォーカス量と第1焦点検出信号と第2焦点検出信号間の像ずれ量の関係を示す図(A) is the figure which showed the correspondence of the image pick-up element of embodiment, and pupil division, (b) is the relationship of the image shift | offset | difference amount between a defocus amount, a 1st focus detection signal, and a 2nd focus detection signal in embodiment. Figure showing 本発明の実施形態におけるデフォーカス量算出処理を説明するためのフローチャートFlowchart for explaining defocus amount calculation processing in an embodiment of the present invention 結像光学系の射出瞳距離と撮像素子の設定瞳距離との差が像高の大きな画素における瞳分割に与える影響を模式的に示した図Diagram showing the effect of the difference between the exit pupil distance of the imaging optical system and the set pupil distance of the image sensor on the pupil division for pixels with a large image height 本発明の実施形態において撮像信号の合焦位置で撮像素子の周辺部の画素によって得られる焦点検出信号の例を示す図The figure which shows the example of the focus detection signal obtained by the pixel of the peripheral part of an image pick-up element in the focus position of an image pick-up signal in embodiment of this invention 焦点検出用画素について算出されたデフォーカス量と、撮像画素におけるデフォーカス量のずれを説明するための図The figure for demonstrating the shift | offset | difference of the defocus amount calculated about the pixel for focus detection, and the defocus amount in an imaging pixel 本発明の実施形態における焦点調節処理を説明するためのフローチャートFlowchart for explaining a focus adjustment process in an embodiment of the present invention

以下、本発明の例示的な実施形態を、図面に基づいて詳細に説明する。
[全体構成]
図1は、本発明の実施形態に係る撮像装置の一例としてのデジタルスチルカメラ100(以下、単にカメラ100という)の機能構成例を示す図である。
Hereinafter, exemplary embodiments of the present invention will be described in detail with reference to the drawings.
[overall structure]
FIG. 1 is a diagram illustrating a functional configuration example of a digital still camera 100 (hereinafter simply referred to as a camera 100) as an example of an imaging apparatus according to an embodiment of the present invention.

第1レンズ群101は撮影光学系(結像光学系)の先端に配置され、光軸に沿って前後に移動可能に保持される。シャッタ102は、静止画撮像時の露光時間を制御するためのシャッタとしてだけでなく、開口径を調節することで撮像時の光量調節を行なう絞りとしても機能する。シャッタ102の背面(撮像素子側)に配置された第2レンズ群103は、シャッタ102と一体となって光軸に沿って前後に可能であり、第1レンズ群101とともにズーム機能を実現する。   The first lens group 101 is disposed at the tip of the photographing optical system (imaging optical system) and is held so as to be movable back and forth along the optical axis. The shutter 102 functions not only as a shutter for controlling the exposure time at the time of still image shooting, but also as an aperture for adjusting the amount of light at the time of shooting by adjusting the aperture diameter. The second lens group 103 disposed on the back surface (on the imaging element side) of the shutter 102 can be moved back and forth along the optical axis integrally with the shutter 102, and realizes a zoom function together with the first lens group 101.

第3レンズ群105はフォーカスレンズであり、光軸に沿って前後に移動可能である。光学ローパスフィルタ106は、撮像素子107の前方に配置され、撮像画像に発生する偽色やモアレを軽減する。撮像素子107は2次元CMOSイメージセンサとその周辺回路で構成される。本実施形態において、撮像素子107は、横方向にm(>1)個、縦方向にn(>1)個の複数の受光素子(画素)が2次元配列され、受光素子にベイヤー配列の原色カラーモザイクフィルタが形成された、2次元単板カラーイメージセンサである。カラーフィルタは受光素子に入射する透過光の波長を画素単位で制限する。   The third lens group 105 is a focus lens and can move back and forth along the optical axis. The optical low-pass filter 106 is disposed in front of the image sensor 107 and reduces false colors and moire generated in the captured image. The image sensor 107 includes a two-dimensional CMOS image sensor and its peripheral circuits. In the present embodiment, the image sensor 107 has a two-dimensional array of a plurality of light receiving elements (pixels) of m (> 1) in the horizontal direction and n (> 1) in the vertical direction. This is a two-dimensional single-plate color image sensor in which a color mosaic filter is formed. The color filter limits the wavelength of transmitted light incident on the light receiving element in units of pixels.

ズームアクチュエータ111は、ズーム駆動回路129の制御に従い、不図示のカム筒を回動して第1レンズ群101と第3レンズ群105の少なくとも一方を光軸に沿って駆動して、ズーム(変倍)機能を実現する。シャッタアクチュエータ112は、シャッタ駆動回路128の制御に従い、シャッタ102開口径を制御して撮像光量を調節すると共に、静止画撮像時の露光時間を制御する。
フォーカスアクチュエータ114は、フォーカス駆動回路126の制御に従い、第3レンズ群105を光軸に沿って駆動する。
Under the control of the zoom drive circuit 129, the zoom actuator 111 rotates a cam cylinder (not shown) to drive at least one of the first lens group 101 and the third lens group 105 along the optical axis to zoom (change). Double) to realize the function. The shutter actuator 112 controls the exposure light amount at the time of capturing a still image while controlling the aperture diameter of the shutter 102 to adjust the amount of imaged light according to the control of the shutter drive circuit 128.
The focus actuator 114 drives the third lens group 105 along the optical axis according to the control of the focus drive circuit 126.

フラッシュ115は、好ましくはキセノン管を用いた閃光照明装置であるが、連続発光するLEDを備えた照明装置であってもよい。AF補助光出力部116は、所定の開口パターンを有するマスクの像を投稿レンズを介して被写界に投影し、低輝度の被写体や低コントラストの被写体に対する焦点検出能力を向上させる。   The flash 115 is preferably a flash illumination device using a xenon tube, but may be an illumination device including LEDs that emit light continuously. The AF auxiliary light output unit 116 projects an image of a mask having a predetermined aperture pattern onto a subject field via a posting lens, and improves the focus detection capability for a low-luminance subject or a low-contrast subject.

CPU121は、カメラ100全体の動作を制御し、図示しない演算部、ROM、RAM、A/Dコンバータ、D/Aコンバータ、通信インターフェイス回路等を有する。CPU121は、ROMに記憶されたプログラムを実行して、カメラ100が有する各種回路を制御し、AF、AE、画像処理、記録等、カメラ100の機能を実現する。   The CPU 121 controls the operation of the entire camera 100, and includes a calculation unit, a ROM, a RAM, an A / D converter, a D / A converter, a communication interface circuit, and the like (not shown). The CPU 121 executes a program stored in the ROM and controls various circuits included in the camera 100 to realize functions of the camera 100 such as AF, AE, image processing, and recording.

また、CPU121には、後述の撮像素子の出力信号を用いた自動焦点検出で必要となる補正値算出係数が記憶されている。
この補正値算出係数は、
結像光学系に関する情報である、
・フォーカス状態(第3レンズ群105の位置もしくは現在の合焦距離)、
・ズーム状態(第1レンズ群101および第2レンズ群103の位置、焦点距離または画角)、
・結像光学系のF値(絞り値)と、
撮像素子に関する情報である、
・撮像素子の設定瞳距離、
・撮像素子の画素サイズ(画素ピッチ)、
との組み合わせごとに複数用意されている。焦点検出時には、結像光学系の状態(フォーカス状態、ズーム状態)と絞り値、撮像素子の設定瞳距離、画素サイズの組み合わせに応じた補正値算出係数が選択され、選択された補正値算出係数と、焦点検出領域または画素の像高とから補正値が算出される。
Further, the CPU 121 stores a correction value calculation coefficient necessary for automatic focus detection using an output signal of an image sensor described later.
This correction value calculation coefficient is
Information about the imaging optical system,
The focus state (the position of the third lens group 105 or the current focusing distance),
Zoom state (position of first lens group 101 and second lens group 103, focal length or angle of view),
-F value (aperture value) of the imaging optical system,
Information about the image sensor
・ Setting pupil distance of image sensor,
-Image sensor pixel size (pixel pitch),
There are several available for each combination. At the time of focus detection, a correction value calculation coefficient is selected according to a combination of the imaging optical system state (focus state, zoom state) and aperture value, image sensor setting pupil distance, and pixel size, and the selected correction value calculation coefficient Then, a correction value is calculated from the focus detection area or the image height of the pixel.

なお、本実施形態では、補正値算出係数をCPU121に記憶する構成としたが、他の構成要素に記憶してもよい。例えば、レンズ交換式の撮像装置においては、結像光学系を有する交換レンズが有する不揮発性メモリに補正値算出係数を記憶してもよい。この場合には、結像光学系の状態および絞り値に応じて、撮像素子の設定瞳距離と画素サイズの複数の組み合わせに対する補正値算出係数を撮像装置に送信する。そして、撮像装置が、例えばCPU121に記憶された撮像素子の設定瞳距離、画素サイズの組み合わせに対応する補正値算出係数を選択すればよい。   In the present embodiment, the correction value calculation coefficient is stored in the CPU 121, but may be stored in another component. For example, in an interchangeable lens imaging device, the correction value calculation coefficient may be stored in a nonvolatile memory included in an interchangeable lens having an imaging optical system. In this case, according to the state of the imaging optical system and the aperture value, correction value calculation coefficients for a plurality of combinations of the pupil distance set for the image sensor and the pixel size are transmitted to the imaging apparatus. Then, the imaging device may select a correction value calculation coefficient corresponding to a combination of the set pupil distance and the pixel size of the imaging element stored in the CPU 121, for example.

フラッシュ制御回路122は、撮像動作に同期してフラッシュ115を点灯制御する。補助光駆動制御回路123は、焦点検出動作時にAF補助光出力部116を点灯制御する。撮像素子駆動回路124は、撮像素子107の動作を制御するとともに、撮像素子107から読み出した画像信号をA/D変換してCPU121に出力する。画像処理回路125は、画像信号に対してγ変換、色補間、JPEG符号化などの画像処理を適用する。   The flash control circuit 122 controls the lighting of the flash 115 in synchronization with the imaging operation. The auxiliary light drive control circuit 123 controls lighting of the AF auxiliary light output unit 116 during the focus detection operation. The image sensor driving circuit 124 controls the operation of the image sensor 107 and A / D converts the image signal read from the image sensor 107 and outputs the image signal to the CPU 121. The image processing circuit 125 applies image processing such as γ conversion, color interpolation, and JPEG encoding to the image signal.

フォーカス駆動回路126は、焦点検出結果に基づいてフォーカスアクチュエータ114を駆動することにより第3レンズ群105を光軸に沿って移動させ、焦点調節を行なう。シャッタ駆動回路128は、シャッタアクチュエータ112を駆動してシャッタ102の開口径及び開閉タイミングを制御する。ズーム駆動回路129は、例えば操作スイッチ132に含まれるズーム操作スイッチの押下によって撮像者から入力されるズーム操作に応じてズームアクチュエータ111を駆動する。   The focus drive circuit 126 moves the third lens group 105 along the optical axis by driving the focus actuator 114 based on the focus detection result, and performs focus adjustment. The shutter drive circuit 128 controls the opening diameter and opening / closing timing of the shutter 102 by driving the shutter actuator 112. The zoom drive circuit 129 drives the zoom actuator 111 according to a zoom operation input from the photographer by pressing a zoom operation switch included in the operation switch 132, for example.

表示器131はLCD等であり、カメラ100の撮像モードに関する情報、撮像前のプレビュー画像と撮像後の確認用画像、焦点検出時の合焦状態の情報等を表示する。操作スイッチ132は、電源スイッチ、レリーズ(撮像トリガ)スイッチ、ズーム操作スイッチ、撮像モード選択スイッチ等を含む。記録媒体133は例えば着脱可能な半導体メモリカードであり、撮像画像を記録する。   The display 131 is an LCD or the like, and displays information related to the imaging mode of the camera 100, a preview image before imaging and a confirmation image after imaging, information on a focusing state at the time of focus detection, and the like. The operation switch 132 includes a power switch, a release (imaging trigger) switch, a zoom operation switch, an imaging mode selection switch, and the like. The recording medium 133 is a detachable semiconductor memory card, for example, and records captured images.

[撮像素子]
図2は、撮像素子107における撮像画素と焦点検出画素の配置例を模式的に示す図であり、撮像画素が横4画素×縦4画素配列された領域を代表的に示している。本実施形態においては、各撮像画素の光電変換領域が横方向に2分割されており、各光電変換領域が焦点検出用画素として機能する。従って、図2においては、焦点検出画素が横8画素×縦4画素配列された領域とも言うことができる。
[Image sensor]
FIG. 2 is a diagram schematically illustrating an arrangement example of the imaging pixels and the focus detection pixels in the imaging element 107, and representatively illustrates an area in which the imaging pixels are arranged in 4 horizontal pixels × 4 vertical pixels. In the present embodiment, the photoelectric conversion area of each imaging pixel is divided into two in the horizontal direction, and each photoelectric conversion area functions as a focus detection pixel. Therefore, in FIG. 2, it can be said that the focus detection pixels are regions in which 8 horizontal pixels × 4 vertical pixels are arranged.

本実施形態において、図2の左上の2×2の画素群200は、撮像素子107に設けられた原色ベイヤー配列のカラーフィルタの繰り返し単位に対応している。従って、R(赤)の分光感度を有する画素200Rが左上に、G(緑)の分光感度を有する画素200Gが右上と左下に、B(青)の分光感度を有する画素200Bが右下に配置されている。また、図2の右上の画素に代表的に示すように、各撮像画素は、横2×縦1に等分割された光電変換部を有しており、左半分の光電変換部が第1焦点検出画素201、右半分の光電変換部が第2焦点検出画素202として利用可能である。撮像画素として利用する場合には、2つの光電変換部で得られた信号を加算した信号を撮像信号として用いる。   In the present embodiment, the 2 × 2 pixel group 200 in the upper left of FIG. 2 corresponds to a repeating unit of a primary color Bayer array color filter provided in the image sensor 107. Accordingly, the pixel 200R having R (red) spectral sensitivity is arranged at the upper left, the pixel 200G having G (green) spectral sensitivity is arranged at the upper right and lower left, and the pixel 200B having B (blue) spectral sensitivity is arranged at the lower right. Has been. Further, as representatively shown in the upper right pixel in FIG. 2, each imaging pixel has a photoelectric conversion unit equally divided into 2 × 1 in the horizontal direction, and the photoelectric conversion unit on the left half is the first focus. The detection pixel 201 and the right half photoelectric conversion unit can be used as the second focus detection pixel 202. When used as an imaging pixel, a signal obtained by adding signals obtained by two photoelectric conversion units is used as an imaging signal.

図2に示した4×4の撮像画素(8×4焦点検出画素)の配列を撮像素子107の撮像面に多数配置することにより、撮像画像を取得しつつ、画面の様々な位置を焦点検出領域として用いた撮像面位相差検出方式の焦点検出を行うことができる。本実施形態では、撮像画素のピッチ(周期)Pが縦横とも4μmで、画素数Nは、横5575×縦3725=約2075万画素であるものとする。また、焦点検出画素の縦方向のピッチPは撮像画素と同じであるが、横方向のピッチPAFは2μmであり、従って焦点検出画素数NAFが横11150×縦3725=約4150万画素であるものとする。 By arranging a large number of 4 × 4 image pickup pixels (8 × 4 focus detection pixels) array shown in FIG. 2 on the image pickup surface of the image sensor 107, focus detection is performed at various positions on the screen while acquiring a picked-up image. It is possible to perform focus detection using the imaging surface phase difference detection method used as the region. In the present embodiment, it is assumed that the pitch (period) P of the imaging pixels is 4 μm both vertically and horizontally, and the number of pixels N is 5575 × vertical 3725 = approximately 20.75 million pixels. Further, the vertical pitch P of the focus detection pixels is the same as that of the imaging pixels, but the horizontal pitch PAF is 2 μm. Therefore, the focus detection pixel number N AF is 11150 × horizontal 3725 = about 41.5 million pixels. It shall be.

図2に示した1つの撮像画素(ここでは200Gとする)を、撮像素子の受光面側(+z側)から見た平面図を図3(a)に示し、図3(a)のa−a断面を−y側から見た断面図を図3(b)に示す。   FIG. 3A shows a plan view of one imaging pixel (here, 200G) shown in FIG. 2 as viewed from the light receiving surface side (+ z side) of the imaging element, and a- in FIG. FIG. 3B shows a cross-sectional view of the a cross section viewed from the −y side.

図3に示すように、本実施形態の画素200Gでは、各画素の受光側に入射光を集光するためのマイクロレンズ305が形成され、x方向にN分割(2分割)、y方向にN分割(1分割)された光電変換部301と光電変換部302が形成される。光電変換部301と光電変換部302が、それぞれ、第1焦点検出画素201と第2焦点検出画素202に対応する。 As shown in FIG. 3, in the pixel 200G of this embodiment, a microlens 305 for condensing incident light is formed on the light receiving side of each pixel, and is divided into NH (two divisions) in the x direction and in the y direction. N V division (first division) by photoelectric conversion unit 301 and the photoelectric conversion portion 302 is formed. The photoelectric conversion unit 301 and the photoelectric conversion unit 302 correspond to the first focus detection pixel 201 and the second focus detection pixel 202, respectively.

光電変換部301と光電変換部302は、p型層とn型層の間にイントリンシック層を挟んだpin構造フォトダイオードとしても良いし、必要に応じて、イントリンシック層を省略し、pn接合フォトダイオードとしてもよい。   The photoelectric conversion unit 301 and the photoelectric conversion unit 302 may be a pin structure photodiode in which an intrinsic layer is sandwiched between a p-type layer and an n-type layer, or an intrinsic layer is omitted as necessary, and a pn junction is formed. A photodiode may be used.

各画素には、マイクロレンズ305と、光電変換部301および光電変換部302との間に、カラーフィルタ306が形成される。また、必要に応じて、第1焦点検出画素201と第2焦点検出画素202とでカラーフィルタの分光透過率を変えても良いし、カラーフィルタを省略してもよい。   In each pixel, a color filter 306 is formed between the microlens 305 and the photoelectric conversion unit 301 and the photoelectric conversion unit 302. Further, as necessary, the spectral transmittance of the color filter may be changed between the first focus detection pixel 201 and the second focus detection pixel 202, or the color filter may be omitted.

図3に示した画素200Gに入射した光は、マイクロレンズ305により集光され、カラーフィルタ306で分光されたのち、光電変換部301と光電変換部302で受光される。   Light incident on the pixel 200 </ b> G illustrated in FIG. 3 is collected by the microlens 305, dispersed by the color filter 306, and then received by the photoelectric conversion unit 301 and the photoelectric conversion unit 302.

光電変換部301と光電変換部302では、受光量に応じて電子とホールが対生成し、空乏層で分離された後、負電荷の電子はn型層に蓄積され、ホールは定電圧源(不図示)に接続されたp型層300を通じて撮像素子107の外部へ排出される。   In the photoelectric conversion unit 301 and the photoelectric conversion unit 302, a pair of electrons and holes are generated according to the amount of received light and separated by the depletion layer, and then negatively charged electrons are accumulated in the n-type layer. It is discharged to the outside of the image sensor 107 through the p-type layer 300 connected to the not shown.

光電変換部301と光電変換部302のn型層に蓄積された電子は、転送ゲートを介して、静電容量部(FD)に転送され、電圧信号に変換される。   The electrons accumulated in the n-type layers of the photoelectric conversion unit 301 and the photoelectric conversion unit 302 are transferred to the electrostatic capacitance unit (FD) through the transfer gate and converted into a voltage signal.

図3に示した本実施形態の画素構造と瞳分割との対応関係を示した概略説明図を図4に示す。図4では、射出瞳面の座標軸と対応を取るために、断面図のx軸とy軸を図3に対して反転させている。   FIG. 4 is a schematic explanatory diagram showing the correspondence between the pixel structure of this embodiment shown in FIG. 3 and pupil division. In FIG. 4, in order to correspond to the coordinate axis of the exit pupil plane, the x-axis and y-axis of the cross-sectional view are inverted with respect to FIG.

図4で、第1焦点検出画素201の第1瞳部分領域501は、重心が−x方向に偏心している光電変換部301の受光面と、マイクロレンズ305によって、概ね共役関係になっており、第1焦点検出画素201で受光可能な瞳領域を表している。第1焦点検出画素201の第1瞳部分領域501は、瞳面上で+X側に重心が偏心している。   In FIG. 4, the first pupil partial region 501 of the first focus detection pixel 201 is substantially conjugate with the light receiving surface of the photoelectric conversion unit 301 whose center of gravity is decentered in the −x direction and the microlens 305. A pupil region that can be received by the first focus detection pixel 201 is shown. The first pupil partial region 501 of the first focus detection pixel 201 has an eccentric center of gravity on the + X side on the pupil plane.

図4で、第2焦点検出画素202の第2瞳部分領域502は、重心が+x方向に偏心している光電変換部302の受光面と、マイクロレンズ305によって、概ね共役関係になっており、第2焦点検出画素202で受光可能な瞳領域を表している。第2焦点検出画素202の第2瞳部分領域502は、瞳面上で−X側に重心が偏心している。   In FIG. 4, the second pupil partial region 502 of the second focus detection pixel 202 is substantially conjugate with the light receiving surface of the photoelectric conversion unit 302 whose center of gravity is decentered in the + x direction and the microlens 305. A pupil region that can be received by the bifocal detection pixel 202 is shown. The second pupil partial region 502 of the second focus detection pixel 202 has an eccentric center of gravity on the −X side on the pupil plane.

また、図4で、瞳領域500は、光電変換部301と光電変換部302(第1焦点検出画素201と第2焦点検出画素202)を合わせた、画素200G全体で受光可能な瞳領域である。   In FIG. 4, a pupil region 500 is a pupil region that can receive light in the entire pixel 200 </ b> G including the photoelectric conversion unit 301 and the photoelectric conversion unit 302 (the first focus detection pixel 201 and the second focus detection pixel 202). .

撮像面位相差検出方式の自動焦点検出(以下、撮像面位相差AFという)では、撮像素子に設けられたマイクロレンズを利用して瞳分割するため回折の影響を受ける。図4で、射出瞳面までの瞳距離が数10mmであるのに対し、マイクロレンズの直径は数μmである。そのため、マイクロレンズの絞り値が数万となり、数10mmレベルの回折ボケが生じる。よって、光電変換部301,302の受光面の像は、明瞭な瞳領域や瞳部分領域とはならずに、瞳強度分布(受光率の入射角分布)となる。   In the automatic focus detection of the imaging surface phase difference detection method (hereinafter referred to as imaging surface phase difference AF), pupil division is performed using a microlens provided in the imaging element, and therefore, it is affected by diffraction. In FIG. 4, the pupil distance to the exit pupil plane is several tens of mm, whereas the diameter of the microlens is several μm. For this reason, the aperture value of the microlens becomes several tens of thousands, and diffraction blur of several tens mm level occurs. Therefore, the image of the light receiving surface of the photoelectric conversion units 301 and 302 does not become a clear pupil region or pupil partial region, but becomes a pupil intensity distribution (incident angle distribution of light reception rate).

撮像画素や焦点検出画素のサイズが変わるとマイクロレンズ305などの光学構造が変化するため、画素サイズに応じて瞳強度分布が変化する。なお、「画素サイズ」とは、画素として機能するための構成を一式有する領域のサイズであり、図2の例では、マイクロレンズ1つが設けられている領域のサイズである。従って、本実施形態では撮像画素と焦点検出画素のサイズが同一であり、画素周期または画素ピッチ(図2におけるP)が画素サイズに等しい。例えば、第1焦点検出画素と第2焦点検出画素のそれぞれが撮像画素と同じサイズを有する形態であれば、焦点検出用画素のサイズは撮像画素の2倍となる。   When the size of the imaging pixel or the focus detection pixel changes, the optical structure such as the microlens 305 changes, so that the pupil intensity distribution changes according to the pixel size. The “pixel size” is a size of a region having a set of structures for functioning as a pixel, and in the example of FIG. 2, is a size of a region where one microlens is provided. Therefore, in this embodiment, the size of the imaging pixel and the focus detection pixel is the same, and the pixel period or the pixel pitch (P in FIG. 2) is equal to the pixel size. For example, if each of the first focus detection pixel and the second focus detection pixel has the same size as the imaging pixel, the size of the focus detection pixel is twice that of the imaging pixel.

図4に示した結像光学系の射出瞳面のX軸に沿った焦点検出画素の瞳強度分布の例を、図5に示す。焦点検出画素のサイズが相対的に小さい場合の瞳強度分布の例を実線で、サイズが相対的に大きい場合の瞳強度分布の例を破線で示す。第1焦点検出画素の瞳強度分布と第2焦点検出画素の瞳強度分布を加算したものが撮像画素の瞳強度分布となる。焦点検出画素(撮像画素)のサイズ変化に伴い、瞳強度分布が変化することがわかる。撮像面位相差AFでは、この瞳強度分布の変化に伴い、焦点検出画素(第1焦点検出画素、第2焦点検出画素)が受光する光束および、撮像画素が受光する光束が変化する。   FIG. 5 shows an example of the pupil intensity distribution of the focus detection pixels along the X axis of the exit pupil plane of the imaging optical system shown in FIG. An example of the pupil intensity distribution when the size of the focus detection pixel is relatively small is indicated by a solid line, and an example of the pupil intensity distribution when the size is relatively large is indicated by a broken line. The sum of the pupil intensity distribution of the first focus detection pixel and the pupil intensity distribution of the second focus detection pixel is the pupil intensity distribution of the imaging pixel. It can be seen that the pupil intensity distribution changes with the size change of the focus detection pixel (imaging pixel). In the imaging plane phase difference AF, the light flux received by the focus detection pixels (first focus detection pixel and second focus detection pixel) and the light flux received by the imaging pixels change with the change in the pupil intensity distribution.

本実施形態の撮像素子と瞳分割との対応関係を示した概略図を図6(a)に示す。第1瞳部分領域501と第2瞳部分領域502の異なる瞳部分領域を通過した光束は、撮像素子の各(撮像)画素に、撮像面800からそれぞれ異なる角度で入射し、2×1分割された光電変換部301および302で受光される。なお、本実施形態は、瞳領域が水平方向に2つに瞳分割されているが、必要に応じて、垂直方向に瞳分割を行ってもよい。   FIG. 6A is a schematic diagram showing the correspondence between the image sensor of this embodiment and pupil division. The light beams that have passed through different pupil partial areas of the first pupil partial area 501 and the second pupil partial area 502 are incident on each (imaging) pixel of the imaging element from the imaging plane 800 at different angles and are divided by 2 × 1. The photoelectric conversion units 301 and 302 receive the light. In this embodiment, the pupil region is divided into two pupils in the horizontal direction, but pupil division may be performed in the vertical direction as necessary.

撮像素子107には、結像光学系の第1瞳部分領域501を通過する光束を受光する第1焦点検出画素201と、第1瞳部分領域と異なる結像光学系の第2瞳部分領域502を通過する光束を受光する第2焦点検出画素202を有する撮像画素が配列されている。従って、撮像画素は、結像光学系の第1瞳部分領域501と第2瞳部分領域502を合わせた瞳領域500を通過する光束を受光する。   The imaging element 107 includes a first focus detection pixel 201 that receives a light beam passing through the first pupil partial region 501 of the imaging optical system, and a second pupil partial region 502 of the imaging optical system that is different from the first pupil partial region. An image pickup pixel having a second focus detection pixel 202 that receives a light beam passing through is arranged. Therefore, the imaging pixel receives a light beam passing through the pupil region 500 that is a combination of the first pupil partial region 501 and the second pupil partial region 502 of the imaging optical system.

なお、撮像素子が有する全ての画素が複数の光電変換部を有するのではなく、撮像画素、第1焦点検出画素、第2焦点検出画素を個別の画素構成としてもよい。あるいは、光電変換部を1つ有する撮像画素と、光電変換部を2つ有する(撮像画素としても使用可能な)焦点検出画素とが配置されても良い。   Note that not all of the pixels included in the imaging element have a plurality of photoelectric conversion units, but the imaging pixels, the first focus detection pixels, and the second focus detection pixels may have individual pixel configurations. Alternatively, an imaging pixel having one photoelectric conversion unit and a focus detection pixel having two photoelectric conversion units (which can also be used as an imaging pixel) may be arranged.

本実施形態で、生成手段としてのCPU121は、複数の第1焦点検出画素201で得られる信号から第1焦点信号を生成し、複数の第2焦点検出画素202から得られる信号から第2焦点信号を生成して焦点検出を行う。また、撮像素子の撮像画素ごとに第1焦点検出画素201と第2焦点検出画素202で得られる信号を加算することで、有効画素数Nの解像度の撮像信号(撮像画像)を生成する。   In the present embodiment, the CPU 121 serving as a generation unit generates a first focus signal from signals obtained from the plurality of first focus detection pixels 201, and second focus signals from signals obtained from the plurality of second focus detection pixels 202. To detect the focus. Further, by adding the signals obtained by the first focus detection pixel 201 and the second focus detection pixel 202 for each image pickup pixel of the image pickup device, an image pickup signal (captured image) having a resolution of N effective pixels is generated.

[デフォーカス量と像ずれ量の関係]
以下、本実施形態の撮像素子により取得される第1焦点検出信号と第2焦点検出信号のデフォーカス量と像ずれ量の関係について説明する。
図6(b)に、デフォーカス量と第1焦点検出信号と第2焦点検出信号間の像ずれ量の概略関係図を示す。撮像面800に撮像素子(不図示)が配置され、図4、図6(a)と同様に、結像光学系の射出瞳が、第1瞳部分領域501と第2瞳部分領域502に2分割される。
[Relationship between defocus amount and image shift amount]
Hereinafter, the relationship between the defocus amount and the image shift amount of the first focus detection signal and the second focus detection signal acquired by the image sensor of the present embodiment will be described.
FIG. 6B shows a schematic relationship diagram of the defocus amount and the image shift amount between the first focus detection signal and the second focus detection signal. An imaging element (not shown) is arranged on the imaging surface 800, and the exit pupil of the imaging optical system is divided into two in the first pupil partial region 501 and the second pupil partial region 502, as in FIGS. Divided.

デフォーカス量dの大きさ|d|は、被写体の結像位置から撮像面800までの距離である。また、デフォーカス量dが負(d<0)の場合は、被写体の結像位置が撮像面800より被写体側にある前ピン状態、正(d>0)の場合は、被写体の結像位置が撮像面800より被写体の反対側にある後ピン状態を意味する。そして、被写体の結像位置が撮像面800にある合焦状態で、デフォーカス量dの大きさは0となる。図6(a)で、被写体801は合焦状態(d=0)にあり、被写体802は前ピン状態(d<0)の例を示している。前ピン状態(d<0)と後ピン状態(d>0)を合わせて、デフォーカス状態(|d|>0)と呼ぶ。   The magnitude | d | of the defocus amount d is the distance from the imaging position of the subject to the imaging surface 800. Further, when the defocus amount d is negative (d <0), the imaging position of the subject is a front pin state that is closer to the subject than the imaging surface 800, and when it is positive (d> 0), the imaging position of the subject. Means a rear pin state in which the image pickup surface 800 is on the opposite side of the subject. The defocus amount d is zero when the subject is focused on the imaging surface 800. In FIG. 6A, the subject 801 is in an in-focus state (d = 0), and the subject 802 shows an example of a front pin state (d <0). The front pin state (d <0) and the rear pin state (d> 0) are collectively referred to as a defocus state (| d |> 0).

前ピン状態(d<0)では、被写体802からの光束のうち、第1瞳部分領域501(第2瞳部分領域502)を通過した光束は、撮像面800より被写体側の位置で一度集光する。そして、その後、光束の重心位置G1(G2)を中心として幅Γ1(Γ2)に広がり、撮像面800でボケた像となる。ボケた像は、それを受光する複数の画素のそれぞれで第1焦点検出画素201(第2焦点検出画素202)により電気信号に変換される。そして、第1焦点検出画素201群(第2焦点検出画素202群)の信号から第1焦点検出信号(第2焦点検出信号)が、生成手段としてのCPU121により生成される。よって、第1焦点検出信号(第2焦点検出信号)は、撮像面800上の重心位置G1(G2)に、被写体802が幅Γ1(Γ2)にボケた被写体像として記録される。   In the front pin state (d <0), the luminous flux that has passed through the first pupil partial area 501 (second pupil partial area 502) out of the luminous flux from the subject 802 is once condensed at a position on the subject side from the imaging plane 800. To do. Then, after that, the image expands in the width Γ1 (Γ2) with the center of gravity position G1 (G2) of the light beam as the center, resulting in a blurred image on the imaging surface 800. The blurred image is converted into an electrical signal by the first focus detection pixel 201 (second focus detection pixel 202) in each of the plurality of pixels that receive the blurred image. Then, a first focus detection signal (second focus detection signal) is generated from a signal of the first focus detection pixel 201 group (second focus detection pixel 202 group) by the CPU 121 as a generation unit. Therefore, the first focus detection signal (second focus detection signal) is recorded as a subject image in which the subject 802 is blurred by the width Γ1 (Γ2) at the gravity center position G1 (G2) on the imaging surface 800.

被写体像のボケ幅Γ1(Γ2)は、デフォーカス量dの大きさ|d|の増加に概ね比例して増加する。同様に、第1焦点検出信号と第2焦点検出信号間の被写体像の像ずれ量p(=光束の重心位置の差G1−G2)の大きさ|p|も、デフォーカス量dの大きさ|d|の増加に概ね比例して増加していく。後ピン状態(d>0)の場合、第1焦点検出信号と第2焦点検出信号間の被写体像の像ずれ方向が前ピン状態と反対となることをのぞき、デフォーカス量の大きさ|d|と被写体像のボケ幅、像ずれ量pとの関係は同様である。   The blur width Γ1 (Γ2) of the subject image increases approximately in proportion to the increase of the magnitude | d | of the defocus amount d. Similarly, the magnitude | p | of the object image displacement amount p (= difference G1-G2 in the center of gravity of the light beam) between the first focus detection signal and the second focus detection signal is also the size of the defocus amount d. It increases almost in proportion to the increase of | d |. In the rear pin state (d> 0), the magnitude of the defocus amount | d except that the image shift direction of the subject image between the first focus detection signal and the second focus detection signal is opposite to the front pin state. The relationship between |, the blur width of the subject image, and the image shift amount p is the same.

したがって、第1焦点検出信号と第2焦点検出信号、もしくは、第1焦点検出信号と第2焦点検出信号を加算した撮像信号のデフォーカス量の大きさの増加に伴い、第1焦点検出信号と第2焦点検出信号間の像ずれ量の大きさが増加する。   Accordingly, the first focus detection signal and the second focus detection signal, or the first focus detection signal and the first focus detection signal are increased in accordance with an increase in the defocus amount of the imaging signal obtained by adding the first focus detection signal and the second focus detection signal. The amount of image shift between the second focus detection signals increases.

[焦点検出]
以下、本実施形態における撮像面位相差AFについて説明する。
本実施形態の撮像面位相差AFでは、第1焦点検出信号と第2焦点検出信号を相対的にシフトさせて信号の一致度を表す相関量を計算し、相関(信号の一致度)が良くなるシフト量から像ずれ量を検出する。撮像信号のデフォーカス量が増加すると第1焦点検出信号と第2焦点検出信号間の像ずれ量が増加する関係性から、像ずれ量をデフォーカス量に変換して焦点検出を行う。
[Focus detection]
Hereinafter, the imaging surface phase difference AF in the present embodiment will be described.
In the imaging surface phase difference AF of the present embodiment, the first focus detection signal and the second focus detection signal are relatively shifted to calculate the amount of correlation indicating the degree of coincidence of signals, and the correlation (the degree of coincidence of signals) is good. The image shift amount is detected from the shift amount. Since the image shift amount between the first focus detection signal and the second focus detection signal increases as the defocus amount of the imaging signal increases, the focus detection is performed by converting the image shift amount into the defocus amount.

図7に、本実施形態におけるデフォーカス量算出処理を説明するためのフローチャートである。
S110でCPU121は、撮像素子駆動回路124を通じて、撮像素子の有効画素領域に設定されている焦点検出領域(AF枠)に関連付けられた領域に含まれる複数の画素から信号を読み出す。焦点検出領域は固定であってもよいし、ユーザが操作スイッチ132などを用いて選択可能であってもよい。なお、通常、焦点検出信号を生成するために画素を読み出す領域は、像ずれ検出方向において焦点検出領域よりも大きい。CPU121は、読み出した信号のうち、第1焦点検出画素から得られる複数の信号から第1焦点検出信号を生成し、第2焦点検出画素から得られる複数の信号から第2焦点検出信号を生成する。
FIG. 7 is a flowchart for explaining the defocus amount calculation processing in the present embodiment.
In S110, the CPU 121 reads signals from a plurality of pixels included in an area associated with a focus detection area (AF frame) set as an effective pixel area of the image sensor through the image sensor driving circuit 124. The focus detection area may be fixed, or may be selectable by the user using the operation switch 132 or the like. Note that the area from which pixels are read in order to generate a focus detection signal is usually larger than the focus detection area in the image shift detection direction. The CPU 121 generates a first focus detection signal from a plurality of signals obtained from the first focus detection pixel among the read signals, and generates a second focus detection signal from the plurality of signals obtained from the second focus detection pixel. .

S120でCPU121は信号データ量を抑制するため、画像処理回路125を用い、第1焦点検出信号と第2焦点検出信号のそれぞれに対し、横方向に3画素加算処理を行う。さらに、CPU121は画像処理回路125を用い、RGB信号をY(輝度)信号にするためにベイヤー(RGB)加算処理を行う。これら2つの加算処理を合わせて画素加算処理と呼ぶ。   In S120, the CPU 121 uses the image processing circuit 125 to suppress the signal data amount, and performs a three-pixel addition process in the horizontal direction for each of the first focus detection signal and the second focus detection signal. Further, the CPU 121 uses the image processing circuit 125 to perform a Bayer (RGB) addition process in order to convert the RGB signal into a Y (luminance) signal. These two addition processes are collectively referred to as a pixel addition process.

S130でCPU121は、画素加算処理後の第1焦点検出信号と第2焦点検出信号に、それぞれ、シェーディング補正処理(光学補正処理)を行う。
ここで、第1焦点検出信号と第2焦点検出信号の瞳ずれによるシェーディングについて説明する。図8は、結像光学系の射出瞳距離Dlと撮像素子の設定瞳距離Dsとの差が像高の大きな(光軸から離れて位置する)画素における瞳分割に与える影響を模式的に示した図である。
In S130, the CPU 121 performs a shading correction process (optical correction process) on the first focus detection signal and the second focus detection signal after the pixel addition process, respectively.
Here, the shading by the pupil shift of the first focus detection signal and the second focus detection signal will be described. FIG. 8 schematically shows the influence of the difference between the exit pupil distance Dl of the imaging optical system and the set pupil distance Ds of the image sensor on the pupil division in a pixel having a large image height (positioned away from the optical axis). It is a figure.

図8(a)は、結像光学系の射出瞳距離Dlと撮像素子107の設定瞳距離Dsが同じ場合を示している。この場合は、像高が小さい(光軸近くに位置する)画素および像高が大きい(光軸から離れて位置する)画素のいずれでも、第1瞳部分領域501と第2瞳部分領域502により、結像光学系の射出瞳400が、概ね均等に瞳分割される。   FIG. 8A shows a case where the exit pupil distance Dl of the imaging optical system and the set pupil distance Ds of the image sensor 107 are the same. In this case, the first pupil partial region 501 and the second pupil partial region 502 are used for both a pixel having a small image height (positioned near the optical axis) and a pixel having a large image height (positioned away from the optical axis). The exit pupil 400 of the imaging optical system is divided into pupils almost uniformly.

図8(b)は、結像光学系の射出瞳距離Dlが撮像素子の設定瞳距離Dsより短い(D1<Ds)場合を示している。この場合、撮像素子の周辺部の画素では、結像光学系の射出瞳400と撮像素子107の入射瞳とにずれが生じ、結像光学系の射出瞳400が、平面図で示すように不均一に分割されてしまう。
図8(c)は、結像光学系の射出瞳距離Dlが撮像素子の設定瞳距離Dsより長い(D1>Ds)場合を示している。この場合も、D1<Dsの場合と同様、撮像素子の周辺部の画素では、結像光学系の射出瞳400と撮像素子107の入射瞳とにずれが生じ、結像光学系の射出瞳400が不均一に分割されてしまう。
FIG. 8B shows a case where the exit pupil distance Dl of the imaging optical system is shorter than the set pupil distance Ds of the image sensor (D1 <Ds). In this case, in the peripheral pixels of the image sensor, a deviation occurs between the exit pupil 400 of the image forming optical system and the entrance pupil of the image sensor 107, and the exit pupil 400 of the image forming optical system is not as shown in the plan view. It is divided evenly.
FIG. 8C shows a case where the exit pupil distance Dl of the imaging optical system is longer than the set pupil distance Ds of the image sensor (D1> Ds). Also in this case, as in the case of D1 <Ds, in the peripheral pixels of the imaging device, a deviation occurs between the exit pupil 400 of the imaging optical system and the entrance pupil of the imaging device 107, and the exit pupil 400 of the imaging optical system. Will be divided unevenly.

瞳分割が不均一になると、第1焦点検出画素と第2焦点検出画素で得られる信号強度に差が生じるため、第1焦点検出信号と第2焦点検出信号の強度が不均一となる(一方の強度が大きくなり、他方の強度が小さくなる)シェーディングが生じる。   If pupil division becomes non-uniform, there will be a difference in signal intensity obtained between the first focus detection pixel and the second focus detection pixel, so the intensity of the first focus detection signal and the second focus detection signal will be non-uniform (one side) ) Increases in intensity and decreases in intensity on the other side).

S130でCPU121は、焦点検出領域の像高と、撮像レンズ(結像光学系)の絞り値(F値)と、射出瞳距離とに応じて、第1焦点検出信号の第1シェーディング補正係数と、第2焦点検出信号の第2シェーディング補正係数を、それぞれ生成する。そして、CPU121は、第1シェーディング補正係数を第1焦点検出信号に乗算し、第2シェーディング補正係数を第2焦点検出信号に乗算して、第1焦点検出信号と第2焦点検出信号のシェーディング補正処理(光学補正処理)を行う。焦点検出領域の像高は、焦点検出領域に含まれる画素位置の代表的な像高であってよく、例えば中心位置の像高であってよい。   In S130, the CPU 121 determines the first shading correction coefficient of the first focus detection signal according to the image height of the focus detection area, the aperture value (F value) of the imaging lens (imaging optical system), and the exit pupil distance. The second shading correction coefficient of the second focus detection signal is generated respectively. Then, the CPU 121 multiplies the first focus detection signal by the first shading correction coefficient, and multiplies the second focus detection signal by the second shading correction coefficient, thereby shading correction of the first focus detection signal and the second focus detection signal. Processing (optical correction processing) is performed. The image height of the focus detection area may be a representative image height of a pixel position included in the focus detection area, for example, an image height of the center position.

第1焦点検出信号と第2焦点検出信号との相関(信号の一致度)を基に、デフォーカス量を算出する際、上述のシェーディングが生じると、デフォーカス量の精度が低下する場合がある。そのため、本実施形態では、焦点検出信号にシェーディング補正を行うことで、精度の良いデフォーカス量の算出を実現する。   When calculating the defocus amount based on the correlation (signal coincidence) between the first focus detection signal and the second focus detection signal, if the above-described shading occurs, the accuracy of the defocus amount may decrease. . Therefore, in this embodiment, accurate calculation of the defocus amount is realized by performing shading correction on the focus detection signal.

なお、シェーディングが生じる原因として、撮像素子107の設定瞳距離Dsは変化せず、結像光学系の射出瞳距離D1が変化する場合を説明したが、結像光学系の射出瞳距離D1が変化せず、撮像素子107の設定瞳距離Dsが変化する場合も同様である。撮像面位相差AFでは、撮像素子107の設定瞳距離Dsの変化に伴い、焦点検出画素(第1焦点検出画素および第2焦点検出画素)が受光する光束と、撮像画素が受光する光束が変化する。   In addition, although the setting pupil distance Ds of the image pick-up element 107 did not change and the exit pupil distance D1 of the imaging optical system changed as a cause of occurrence of shading, the exit pupil distance D1 of the imaging optical system changed. The same applies to the case where the set pupil distance Ds of the image sensor 107 changes. In the imaging plane phase difference AF, the light flux received by the focus detection pixels (the first focus detection pixel and the second focus detection pixel) and the light flux received by the imaging pixels change with a change in the set pupil distance Ds of the image sensor 107. To do.

図7のS140でCPU121は、精度の良いデフォーカス量を算出するために、第1焦点検出信号と第2焦点検出信号に、特定の通過周波数帯域を有するバンドパスフィルタ処理を行い、信号の一致度を改善する。バンドパスフィルタの例としては、DC成分をカットしてエッジ抽出を行う{1,4,4,4,0,−4,−4,−4,−1}などの差分型フィルタや、高周波ノイズ成分を抑制する{1,2,1}などの加算型フィルタがある。なお、これらは空間フィルタの係数列である。   In S140 of FIG. 7, the CPU 121 performs bandpass filter processing having a specific pass frequency band on the first focus detection signal and the second focus detection signal in order to calculate a precise defocus amount, thereby matching the signals. Improve the degree. Examples of bandpass filters include differential filters such as {1, 4, 4, 4, 0, -4, -4, -4, -1} that perform edge extraction by cutting DC components, and high-frequency noise. There are additive filters such as {1, 2, 1} that suppress components. Note that these are coefficient strings of the spatial filter.

次に、図7のS150で第1取得手段としてのCPU121は、フィルタ処理後の第1焦点検出信号と第2焦点検出信号を相対的に瞳分割方向にシフトさせるシフト処理を行い、信号の一致度を表す相関量を算出する。   Next, in S150 of FIG. 7, the CPU 121 as the first acquisition unit performs a shift process for relatively shifting the first focus detection signal and the second focus detection signal after the filter process in the pupil division direction, thereby matching the signals. A correlation amount representing the degree is calculated.

フィルタ処理後の第1焦点検出信号および第2焦点検出信号を構成するk番目の信号をそれぞれA(k),B(k)、第1焦点検出信号(および第2焦点検出信号)を構成する信号の数をWとする。従って、番号kの範囲は1〜Wである。シフト処理によるシフト量をs、シフト量sの範囲をΓとして、相関量CORは、式(1)により算出される。

Figure 0006239857
The kth signals constituting the first focus detection signal and the second focus detection signal after the filter processing constitute A (k), B (k), and the first focus detection signal (and the second focus detection signal), respectively. Let W be the number of signals. Therefore, the range of the number k is 1 to W. The correlation amount COR is calculated by the equation (1), where s is the shift amount by the shift process and Γ is the range of the shift amount s.
Figure 0006239857

シフト量sのシフト処理により、CPU121は、第1焦点検出信号を構成するk番目の信号A(k)と第2焦点検出信号を構成する(k−s)番目の信号B(k−s)とを減算し、シフト減算信号を生成する。そしてCPU121は生成したシフト減算信号の絶対値を計算し、焦点検出領域に対応する範囲W内でkの値を順次変えながら累積して、シフト量sに対する相関量COR(s)を算出する。同じシフト量sで異なる複数の画素行について算出された複数の相関量を加算して相関量COR(s)を算出してもよい。   By the shift process of the shift amount s, the CPU 121 causes the kth signal A (k) constituting the first focus detection signal and the (ks) th signal B (ks) constituting the second focus detection signal. Are subtracted to generate a shift subtraction signal. Then, the CPU 121 calculates the absolute value of the generated shift subtraction signal, accumulates it while sequentially changing the value of k within the range W corresponding to the focus detection area, and calculates the correlation amount COR (s) with respect to the shift amount s. The correlation amount COR (s) may be calculated by adding a plurality of correlation amounts calculated for a plurality of different pixel rows with the same shift amount s.

S160で第1取得手段としてのCPU121は、相関量CORから、サブピクセル演算により、相関量が最小値となる実数値のシフト量を算出して像ずれ量pとする。像ずれ量pに、焦点検出領域の像高と、撮像レンズ(結像光学系)の絞り値と、射出瞳距離とに応じた変換係数Kを乗じたpKをデフォーカス量(Def)として算出する。   In S160, the CPU 121 as the first acquisition unit calculates a real-valued shift amount at which the correlation amount is the minimum value from the correlation amount COR by sub-pixel calculation, and sets it as the image shift amount p. The defocus amount (Def) is calculated by multiplying the image shift amount p by the conversion factor K corresponding to the image height of the focus detection region, the aperture value of the imaging lens (imaging optical system), and the exit pupil distance. To do.

焦点検出画素(第1焦点検出画素、第2焦点検出画素)が受光する光束と、撮像画素が受光する光束とが異なるため、結像光学系の各収差(球面収差、非点収差、コマ収差など)が与える影響が焦点検出信号と撮像信号とで異なる。そのため、上述した方法で撮像面位相差AFによって算出したデフォーカス量Defが0となる焦点検出信号の合焦位置と、本来検出したい撮像信号の合焦位置(撮像信号のMTFピーク位置)との間に差が生じる場合がある。   Since the light beam received by the focus detection pixel (the first focus detection pixel and the second focus detection pixel) is different from the light beam received by the imaging pixel, each aberration of the imaging optical system (spherical aberration, astigmatism, coma aberration) And the like are different between the focus detection signal and the imaging signal. Therefore, the focus position of the focus detection signal where the defocus amount Def calculated by the imaging surface phase difference AF by the above-described method is 0 and the focus position of the imaging signal to be originally detected (the MTF peak position of the imaging signal). There may be a difference between them.

図9に、本実施形態の撮像素子の周辺部の画素において、撮像信号の合焦位置で得られる第1焦点検出信号(破線)と第2焦点検出信号(実線)の例を示す。撮像信号については合焦位置であるが、結像光学系の収差の影響により、第1焦点検出信号と第2焦点検出信号の形状が異なっている。そのため、図9の例では、撮像信号の合焦位置で、第1焦点検出信号と第2焦点検出信号間の像ずれ量pが0にならない。つまり、撮像面位相差AFで得られる、第1焦点検出信号と第2焦点検出信号との像ずれ量pが0になる合焦位置は、撮像信号の合焦位置と一致せず、差が生じる。   FIG. 9 shows an example of the first focus detection signal (broken line) and the second focus detection signal (solid line) obtained at the focus position of the image pickup signal in the peripheral pixels of the image sensor of the present embodiment. Although the imaging signal is the in-focus position, the shapes of the first focus detection signal and the second focus detection signal are different due to the influence of the aberration of the imaging optical system. Therefore, in the example of FIG. 9, the image shift amount p between the first focus detection signal and the second focus detection signal does not become zero at the focus position of the imaging signal. That is, the in-focus position where the image shift amount p between the first focus detection signal and the second focus detection signal obtained by the imaging surface phase difference AF is 0 does not match the in-focus position of the imaging signal, and the difference is Arise.

図10において、破線は撮像面位相差AFで得られたデフォーカス量(検出デフォーカス量)の例を示す。図10の横軸は、設定デフォーカス量であり、縦軸は検出デフォーカス量である。図9に示した第1焦点検出信号と第2焦点検出信号は、図10の設定デフォーカス量0[mm]における第1焦点検出信号と第2焦点検出信号である。設定デフォーカス量0(=撮像信号の合焦位置)において、検出デフォーカス量が後ピン側に約50μmオフセットしており、検出デフォーカス量に基づく合焦位置と、撮像信号の合焦位置との間に約50μmの差異が生じていることがわかる。   In FIG. 10, a broken line indicates an example of the defocus amount (detected defocus amount) obtained by the imaging surface phase difference AF. The horizontal axis in FIG. 10 is the set defocus amount, and the vertical axis is the detected defocus amount. The first focus detection signal and the second focus detection signal shown in FIG. 9 are the first focus detection signal and the second focus detection signal at the set defocus amount 0 [mm] in FIG. When the set defocus amount is 0 (= focus position of the imaging signal), the detected defocus amount is offset by about 50 μm toward the rear pin, and the focus position based on the detected defocus amount, the focus position of the imaging signal, and It can be seen that there is a difference of about 50 μm between the two.

このような、結像光学系の収差による、検出デフォーカス量に基づく合焦位置と本来検出したい合焦位置とのずれΔDefを補正し、撮像面位相差AFによる高精度な焦点検出を可能とするため、本実施形態では検出デフォーカス量を補正する。   Such a deviation ΔDef between the in-focus position based on the detected defocus amount and the in-focus position to be originally detected due to the aberration of the imaging optical system can be corrected, and high-precision focus detection by the imaging surface phase difference AF becomes possible. Therefore, in this embodiment, the detected defocus amount is corrected.

撮像面位相差AFで算出される検出デフォーカス量に基づく合焦位置と、撮像信号の合焦位置との差ΔDefは、焦点検出画素(第1焦点検出画素、第2焦点検出画素)が受光する光束と、撮像画素が受光する光束とが異なることに起因して生じる。そして、これらの差異は、
(1)結像光学系の焦点調節状態(フォーカス状態、ズーム状態)、
(2)結像光学系の絞り値、
(3)焦点検出領域の像高、
(4)撮像素子の設定瞳距離、
(5)撮像画素、第1焦点検出画素、第2焦点検出画素の画素サイズ
の組み合わせに応じて変化する。(1)、(2)は結像光学系に関する情報であり、(4)、(5)は撮像素子に関する情報である。
The focus detection pixel (first focus detection pixel, second focus detection pixel) receives the difference ΔDef between the focus position based on the detected defocus amount calculated by the imaging surface phase difference AF and the focus position of the imaging signal. This is caused by the difference between the luminous flux to be received and the luminous flux received by the imaging pixel. And these differences are
(1) Focus adjustment state (focus state, zoom state) of the imaging optical system,
(2) Aperture value of the imaging optical system,
(3) the image height of the focus detection area,
(4) set pupil distance of image sensor,
(5) Changes in accordance with a combination of pixel sizes of the imaging pixel, the first focus detection pixel, and the second focus detection pixel. (1) and (2) are information relating to the imaging optical system, and (4) and (5) are information relating to the image sensor.

したがって、本実施形態では、これら条件(1)〜(5)の組み合わせに応じた補正値(上述したΔDefに等しい)を算出し、検出デフォーカス量を補正値によって補正する。なお、合焦位置のずれに影響する条件(1)〜(5)のうち、本実施形態のように(5)における焦点検出画素と撮像画像のサイズが等しい場合には、いずれかのサイズであってよい。画素サイズが異なる場合には、サイズの組み合わせに応じた補正値を用いる。   Therefore, in the present embodiment, a correction value (equivalent to ΔDef described above) corresponding to the combination of these conditions (1) to (5) is calculated, and the detected defocus amount is corrected by the correction value. Of the conditions (1) to (5) that affect the shift of the in-focus position, if the size of the focus detection pixel in (5) is the same as that of the captured image as in this embodiment, any size is used. It may be. When the pixel sizes are different, a correction value corresponding to a combination of sizes is used.

[焦点調節]
本実施形態における焦点調節処理を、図11に示すフローチャートを用いて説明する。
S100では、CPU121などにより、図7を用いて上述したように検出デフォーカス量(Def)を算出する。
[Focus adjustment]
The focus adjustment process in this embodiment will be described with reference to the flowchart shown in FIG.
In S100, the detected defocus amount (Def) is calculated by the CPU 121 and the like as described above with reference to FIG.

S200で第2取得手段としてのCPU121は、結像光学系の焦点調節状態と、結像光学系の絞り値と、撮像素子の像高と、撮像素子の設定瞳距離と、画素サイズとの組み合わせに応じた補正値ΔDefを算出する。   In S200, the CPU 121 as the second acquisition unit combines the focus adjustment state of the imaging optical system, the aperture value of the imaging optical system, the image height of the imaging device, the set pupil distance of the imaging device, and the pixel size. A correction value ΔDef according to the above is calculated.

まずCPU121は、現在の複結像光学系のフォーカス状態FS、ズーム状態ZS、絞り値F、撮像素子の設定瞳距離D、画素サイズSの組み合わせに対応した6つの補正値算出係数
(FS,ZS,F,D,S)
x2(FS,ZS,F,D,S)
y2(FS,ZS,F,D,S)
x4(FS,ZS,F,D,S)
x2y2(FS,ZS,F,D,S)
y4(FS,ZS,F,D,S)
を選択する。
First, the CPU 121 has six correction value calculation coefficients C 0 (FS) corresponding to combinations of the focus state FS, zoom state ZS, aperture value F, imaging element setting pupil distance D, and pixel size S of the current multi-imaging optical system. , ZS, F, D, S)
C x2 (FS, ZS, F, D, S)
C y2 (FS, ZS, F, D, S)
C x4 (FS, ZS, F, D, S)
C x2y2 (FS, ZS, F, D, S)
C y4 (FS, ZS, F, D, S)
Select.

補正値算出係数の選択は、これら5つのパラメータの複数の組み合わせに対して予め用意され、CPU121などが有する不揮発性記憶装置に記憶された補正値算出係数のテーブルを参照することによって実施することができる。なお、上述したように、5つのパラメータの値のうち、結像光学系に関する情報(FS,ZS,F)は交換レンズから取得する。また、テーブルは交換レンズに記憶しておくこともでき、この場合、交換レンズが有する制御部が、テーブルから現在の結像光学系のパラメータ値と、複数のD,Sとの組み合わせに対応するテーブルを撮像装置へ供給する。そして、撮像装置が、自身の撮像素子のパラメータ値に対応する係数を選択する。なお、記憶容量を節約するため、5つのパラメータの離散値の組み合わせについてのみ補正値算出係数を用意し、現在のパラメータ値に合致する組み合わせがなければ、近い組み合わせに対応する複数の補正値算出係数を補間して補正値算出係数を取得してもよい。   The correction value calculation coefficient is selected by referring to a table of correction value calculation coefficients prepared in advance for a plurality of combinations of these five parameters and stored in the nonvolatile storage device of the CPU 121 or the like. it can. As described above, of the five parameter values, information (FS, ZS, F) related to the imaging optical system is acquired from the interchangeable lens. The table can also be stored in the interchangeable lens. In this case, the control unit included in the interchangeable lens corresponds to the combination of the parameter value of the current imaging optical system from the table and a plurality of D and S. The table is supplied to the imaging device. Then, the imaging device selects a coefficient corresponding to the parameter value of its own imaging device. In order to save storage capacity, correction value calculation coefficients are prepared only for combinations of discrete values of five parameters. If there is no combination that matches the current parameter value, a plurality of correction value calculation coefficients corresponding to the closest combination are prepared. May be interpolated to obtain the correction value calculation coefficient.

次に第2取得手段としてのCPU121は、焦点検出信号の生成に用いた画素群の像高の代表値として、位置焦点検出領域の像高(x,y)を用いて、補正値ΔDef(FS,ZS,F,D,S:x,y)を、式(2)により算出する。

Figure 0006239857
上述の6つの補正値算出係数は式(2)の係数である。式(2)は、AFセンサと2次結像光学系を用いた位相差検出方式の自動焦点検出において、デフォーカス量から得られる合焦位置と実際の合焦位置との差を像高に応じて補正するための補正値を算出する多項式と同様の多項式である。なお、像高からデフォーカス量の補正値を取得する方法は、このような多項式を利用する方法に限定されない。 Next, the CPU 121 as the second acquisition unit uses the image height (x, y) of the position focus detection region as the representative value of the image height of the pixel group used for generating the focus detection signal, and uses the correction value ΔDef (FS , ZS, F, D, S: x, y) are calculated by the equation (2).
Figure 0006239857
The six correction value calculation coefficients described above are the coefficients of Expression (2). Equation (2) is the difference between the focus position obtained from the defocus amount and the actual focus position in the image height in the automatic focus detection of the phase difference detection method using the AF sensor and the secondary imaging optical system. It is a polynomial similar to a polynomial for calculating a correction value for correction accordingly. Note that the method for obtaining the defocus amount correction value from the image height is not limited to the method using such a polynomial.

なお、ここでは、焦点検出領域の像高として、焦点検出領域を代表する画素位置を、光軸と撮像素子とが交わる位置を原点とした直交座標系における画素座標で表した値を用いている。これは、像高の大きさが等しい画素位置であっても、必ずしも補正値が同じとなるとは限らないことによる。また、焦点検出領域の像高として用いる画素位置は、たとえば焦点検出領域の中心位置であってよい。像高は焦点検出領域において1つとする必要は無く、焦点検出信号を生成する画素ラインごとに、焦点検出領域の位相差検出方向における中心位置を像高として用いても良い。   Here, as the image height of the focus detection area, a value representing the pixel position representing the focus detection area in terms of pixel coordinates in an orthogonal coordinate system with the position where the optical axis and the image sensor intersect as the origin is used. . This is because even if the pixel positions have the same image height, the correction values are not always the same. Further, the pixel position used as the image height of the focus detection area may be, for example, the center position of the focus detection area. The image height does not need to be one in the focus detection area, and the center position of the focus detection area in the phase difference detection direction may be used as the image height for each pixel line that generates the focus detection signal.

交換レンズ式の撮像装置では、撮像装置に搭載される撮像素子毎に、撮像素子の構成(設定瞳距離Dと撮像画素の画素サイズS)が異なる。そのため、補正値算出係数を交換レンズ側に記憶させる場合、補正値算出係数を交換レンズの状態(フォーカス状態FS、ズーム状態ZS、絞り値F)ごとに記憶しておくだけでは、撮像素子の構成が異なる場合に正しい補正値を算出できない。したがって、本実施形態では、補正値算出係数を交換レンズの状態(フォーカス状態FS、ズーム状態ZS、絞り値F)と、想定される撮像素子の構成(設定瞳距離Dと撮像画素の画素サイズS)との組み合わせ毎に補正値算出係数を記憶しておく。なお、通常はそうであるが、撮像素子が変化することがなければ、撮像装置側で、交換レンズの機種識別情報ごとに、交換レンズの状態(フォーカス状態FS、ズーム状態ZS、絞り値F)の複数の組み合わせに対して補正値算出係数を記憶しておいてもよい。この場合、撮像装置が交換レンズから機種識別情報と交換レンズの状態(結像光学系に関する情報)を取得し、対応する補正値算出係数を取得することができる。   In an interchangeable lens type imaging device, the configuration of the imaging device (set pupil distance D and pixel size S of the imaging pixel) differs for each imaging device mounted on the imaging device. For this reason, when the correction value calculation coefficient is stored on the interchangeable lens side, the configuration of the image sensor only needs to be stored for each state of the interchangeable lens (focus state FS, zoom state ZS, aperture value F). If the values are different, correct correction values cannot be calculated. Therefore, in the present embodiment, the correction value calculation coefficient is determined based on the state of the interchangeable lens (focus state FS, zoom state ZS, aperture value F), and the assumed configuration of the imaging device (set pupil distance D and pixel size S of the imaging pixel). The correction value calculation coefficient is stored for each combination. Normally, this is the case, but if the image sensor does not change, the state of the interchangeable lens (focus state FS, zoom state ZS, aperture value F) is determined on the image capturing apparatus side for each model identification information of the interchangeable lens. Correction value calculation coefficients may be stored for a plurality of combinations. In this case, the imaging apparatus can acquire the model identification information and the state of the interchangeable lens (information regarding the imaging optical system) from the interchangeable lens, and can obtain the corresponding correction value calculation coefficient.

S300で補正手段としてのCPU121は、補正値ΔDefによりデフォーカス量(Def)を補正し、補正後のデフォーカス量(Def1=Def−ΔDef)を算出する。   In S300, the CPU 121 as the correcting unit corrects the defocus amount (Def) by the correction value ΔDef, and calculates the corrected defocus amount (Def1 = Def−ΔDef).

S400でCPU121は、補正後のデフォーカス量(Def1)の絶対値が予め定めた所定値1より大きい場合は、結像光学系の合焦位置近傍ではないと判定し、処理をS500に進める。S500でCPU121は、フォーカス駆動回路126およびフォーカスアクチュエータを通じて、補正後のデフォーカス量(Def1)に応じた第3レンズ群105(フォーカスレンズ)の駆動を行い、処理をS100に戻す。   If the absolute value of the corrected defocus amount (Def1) is larger than the predetermined value 1 in S400, the CPU 121 determines that the position is not near the in-focus position of the imaging optical system, and advances the process to S500. In S500, the CPU 121 drives the third lens group 105 (focus lens) according to the corrected defocus amount (Def1) through the focus drive circuit 126 and the focus actuator, and returns the process to S100.

一方、S400で補正後のデフォーカス量(Def1)の絶対値が所定値1以下の場合、CPU121は既に結像光学系が合焦位置近傍(フォーカスレンズを駆動する必要が無い)と判定して、焦点調節処理を終了する。   On the other hand, if the absolute value of the defocus amount (Def1) after correction in S400 is less than or equal to the predetermined value 1, the CPU 121 determines that the imaging optical system is already in the vicinity of the in-focus position (no need to drive the focus lens). The focus adjustment process is terminated.

本実施形態によれば、撮像面位相差検出方式によって算出される焦点検出信号のデフォーカス量を、結像光学系の状態に関する情報と、撮像素子に関する情報と、像高との組み合わせに応じた補正値で補正する。そのため、結像光学系の収差によって生じる、デフォーカス量に基づく合焦位置と撮像信号の合焦位置とのずれを軽減し、合焦精度を向上することができる。   According to this embodiment, the defocus amount of the focus detection signal calculated by the imaging surface phase difference detection method depends on the combination of the information about the state of the imaging optical system, the information about the imaging device, and the image height. Correct with the correction value. Therefore, it is possible to reduce the deviation between the focus position based on the defocus amount and the focus position of the imaging signal, which is caused by the aberration of the imaging optical system, and improve the focus accuracy.

(その他の実施形態)
また、本発明は、以下の処理を実行することによっても実現される。即ち、上述した実施形態の機能を実現するソフトウェア(プログラム)を、ネットワーク又は各種記憶媒体を介してシステム或いは装置に供給し、そのシステム或いは装置のコンピュータ(またはCPUやMPU等)がプログラムを読み出して実行する処理である。
(Other embodiments)
The present invention can also be realized by executing the following processing. That is, software (program) that realizes the functions of the above-described embodiments is supplied to a system or apparatus via a network or various storage media, and a computer (or CPU, MPU, or the like) of the system or apparatus reads the program. It is a process to be executed.

Claims (5)

結像光学系の射出瞳の第1部分を通過する光束を受光する複数の第1焦点検出用画素と、前記第1部分と異なる第2部分を通過する光束を受光する複数の第2焦点検出用画素とを有する撮像素子と、
前記撮像素子の焦点検出領域に関連付けられた領域の前記第1焦点検出用画素から得られる信号から第1焦点検出信号を、前記領域の前記第2焦点検出用画素から得られる信号から第2焦点検出信号をそれぞれ生成する第1生成手段と、
前記第1焦点検出信号と前記第2焦点検出信号との像ずれ量から、前記結像光学系のデフォーカス量を取得する第1取得手段と、
前記第1焦点検出用画素から得られる信号と、前記第2焦点検出用画素から得られる信号とを用いて、記録媒体に記録する記録画像を生成する第2生成手段と、
前記結像光学系の状態に関する情報と前記撮像素子に関する情報との組み合わせに応じた補正係数と、前記焦点検出領域の像高とから、前記デフォーカス量の補正値であって、前記第1取得手段にて取得されたデフォーカス量に基づく合焦位置と、前記記録画像の合焦位置との差を補正するための補正値を取得する第2取得手段と、
前記補正値によって前記デフォーカス量を補正し、前記結像光学系の焦点調節に用いるデフォーカス量を取得する補正手段と、
を有することを特徴とする撮像装置。
A plurality of first focus detection pixels for receiving a light beam passing through a first part of an exit pupil of the imaging optical system, and a plurality of second focus detections for receiving a light beam passing through a second part different from the first part. An imaging device having a pixel for use;
The first focus detection signal is obtained from the signal obtained from the first focus detection pixel in the area associated with the focus detection area of the image sensor, and the second focus is obtained from the signal obtained from the second focus detection pixel in the area. First generation means for generating detection signals,
First acquisition means for acquiring a defocus amount of the imaging optical system from an image shift amount between the first focus detection signal and the second focus detection signal;
Second generation means for generating a recording image to be recorded on a recording medium using a signal obtained from the first focus detection pixel and a signal obtained from the second focus detection pixel;
From the correction coefficient according to the combination of the information related to the state of the imaging optical system and the information related to the imaging device, and the image height of the focus detection area, the correction value of the defocus amount , the first acquisition Second acquisition means for acquiring a correction value for correcting a difference between the focus position based on the defocus amount acquired by the means and the focus position of the recorded image ;
Correction means for correcting the defocus amount by the correction value, and obtaining a defocus amount used for focus adjustment of the imaging optical system;
An imaging device comprising:
前記結像光学系に関する情報が、前記結像光学系のフォーカス状態、ズーム状態、および絞り値を含むことを特徴とする請求項1記載の撮像装置。   The imaging apparatus according to claim 1, wherein the information related to the imaging optical system includes a focus state, a zoom state, and an aperture value of the imaging optical system. 前記撮像素子に関する情報が、前記撮像素子の画素サイズおよび設定瞳距離を含むことを特徴とする請求項1または請求項2に記載の撮像装置。   The imaging apparatus according to claim 1, wherein the information regarding the imaging element includes a pixel size and a set pupil distance of the imaging element. 前記第2取得手段は、前記結像光学系の状態に関する情報と前記撮像素子に関する情報との組み合わせに応じて予め記憶された複数の補正係数から、前記撮像装置の有する撮像素子に対応する補正係数を選択することにより、前記デフォーカス量の補正に用いる補正係数を取得することを特徴とする請求項1から請求項3のいずれか1項に記載の撮像装置。   The second acquisition means includes a correction coefficient corresponding to an image sensor included in the image pickup device, based on a plurality of correction coefficients stored in advance according to a combination of information regarding the state of the imaging optical system and information regarding the image sensor. The image pickup apparatus according to claim 1, wherein a correction coefficient used for correcting the defocus amount is acquired by selecting. 結像光学系の射出瞳の第1部分を通過する光束を受光する複数の第1焦点検出用画素と、前記第1部分と異なる第2部分を通過する光束を受光する複数の第2焦点検出用画素とを有する撮像素子を有する撮像装置の制御方法であって、
記撮像素子の焦点検出領域に関連付けられた領域の前記第1焦点検出用画素から得られる信号から第1焦点検出信号を、前記領域の前記第2焦点検出用画素から得られる信号から第2焦点検出信号をそれぞれ生成する第1生成工程と、
記第1焦点検出信号と前記第2焦点検出信号との像ずれ量から、前記結像光学系のデフォーカス量を取得する第1取得工程と、
前記第1焦点検出用画素から得られる信号と、前記第2焦点検出用画素から得られる信号とを用いて、記録媒体に記録する記録画像を生成する第2生成工程と、
前記結像光学系の状態に関する情報と前記撮像素子に関する情報との組み合わせに応じた補正係数と、前記焦点検出領域の像高とから、前記デフォーカス量の補正値であって、前記第1取得工程にて取得されたデフォーカス量に基づく合焦位置と、前記記録画像の合焦位置との差を補正するための補正値を取得する第2取得工程と、
記補正値によって前記デフォーカス量を補正し、前記結像光学系の焦点調節に用いるデフォーカス量を取得する補正工程と、
を有することを特徴とする撮像装置の制御方法。
A plurality of first focus detection pixels for receiving a light beam passing through a first part of an exit pupil of the imaging optical system, and a plurality of second focus detections for receiving a light beam passing through a second part different from the first part. A method for controlling an imaging apparatus having an imaging element having a pixel for use,
The first focus detection signal from the signal obtained from the first focus detection pixels in the region associated with the focus detection area in front Symbol imaging device, first the signal obtained from the second focus detection pixel of the region 2 A first generation step of generating a focus detection signal;
From the image shift amount of the front Symbol the first focus detection signal the second focus detection signal, a first obtaining step of obtaining a defocus amount of the imaging optical system,
A second generation step of generating a recording image to be recorded on a recording medium using a signal obtained from the first focus detection pixel and a signal obtained from the second focus detection pixel;
From the correction coefficient according to the combination of the information related to the state of the imaging optical system and the information related to the imaging device, and the image height of the focus detection area, the correction value of the defocus amount , the first acquisition A second acquisition step of acquiring a correction value for correcting a difference between the focus position based on the defocus amount acquired in the step and the focus position of the recorded image ;
A correction step of the defocus amount is corrected by the previous SL correction value, obtains the defocus amount to be used for focus adjustment of the imaging optical system,
A method for controlling an imaging apparatus, comprising:
JP2013101713A 2013-05-13 2013-05-13 Imaging apparatus and control method thereof Active JP6239857B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2013101713A JP6239857B2 (en) 2013-05-13 2013-05-13 Imaging apparatus and control method thereof

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP2013101713A JP6239857B2 (en) 2013-05-13 2013-05-13 Imaging apparatus and control method thereof

Publications (2)

Publication Number Publication Date
JP2014222291A JP2014222291A (en) 2014-11-27
JP6239857B2 true JP6239857B2 (en) 2017-11-29

Family

ID=52121841

Family Applications (1)

Application Number Title Priority Date Filing Date
JP2013101713A Active JP6239857B2 (en) 2013-05-13 2013-05-13 Imaging apparatus and control method thereof

Country Status (1)

Country Link
JP (1) JP6239857B2 (en)

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6298362B2 (en) * 2014-06-02 2018-03-20 キヤノン株式会社 Imaging apparatus, control method therefor, and imaging system
JP6576114B2 (en) * 2015-06-17 2019-09-18 キヤノン株式会社 IMAGING DEVICE AND IMAGING DEVICE CONTROL METHOD
JP2017049351A (en) * 2015-08-31 2017-03-09 富士通株式会社 Focus position detection device, focus position detection method and focus position detection computer program
JP6682310B2 (en) 2016-03-18 2020-04-15 キヤノン株式会社 Focus detection apparatus and method, and imaging apparatus
JP7022575B2 (en) * 2017-01-13 2022-02-18 キヤノン株式会社 Focus detectors and methods, and imaging devices
US10326926B2 (en) 2017-01-13 2019-06-18 Canon Kabushiki Kaisha Focus detection apparatus and method, and image capturing apparatus
JP7191552B2 (en) 2018-05-30 2022-12-19 キヤノン株式会社 Control device, imaging device, program, and storage medium
JP2019219499A (en) * 2018-06-19 2019-12-26 キヤノン株式会社 Controller, imaging apparatus, control method, program, and storage medium
US11067905B1 (en) 2018-06-29 2021-07-20 Applied Materials, Inc. Real-time autofocus for maskless lithography on substrates
JP7171331B2 (en) * 2018-09-14 2022-11-15 キヤノン株式会社 Imaging device
JP7292909B2 (en) * 2019-03-18 2023-06-19 キヤノン株式会社 Imaging device and focus detection method

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4410804A (en) * 1981-07-13 1983-10-18 Honeywell Inc. Two dimensional image panel with range measurement capability
JP3592147B2 (en) * 1998-08-20 2004-11-24 キヤノン株式会社 Solid-state imaging device
JP3774597B2 (en) * 1999-09-13 2006-05-17 キヤノン株式会社 Imaging device
JP2003241075A (en) * 2002-02-22 2003-08-27 Canon Inc Camera system, camera and photographic lens device
JP2003295047A (en) * 2002-04-05 2003-10-15 Canon Inc Image pickup device and image pickup system
JP2005352043A (en) * 2004-06-09 2005-12-22 Fuji Photo Film Co Ltd Image pickup apparatus and focusing control method
JP4946059B2 (en) * 2006-01-11 2012-06-06 株式会社ニコン Imaging device
JP5212396B2 (en) * 2010-02-10 2013-06-19 株式会社ニコン Focus detection device
JP5901246B2 (en) * 2010-12-13 2016-04-06 キヤノン株式会社 Imaging device
JP5866760B2 (en) * 2010-12-16 2016-02-17 株式会社ニコン Imaging device
KR101777351B1 (en) * 2011-05-16 2017-09-11 삼성전자주식회사 Image pickup device, digital photographing apparatus using the device, auto-focusing method, and computer-readable storage medium for performing the method
JP5784395B2 (en) * 2011-07-13 2015-09-24 オリンパス株式会社 Imaging device

Also Published As

Publication number Publication date
JP2014222291A (en) 2014-11-27

Similar Documents

Publication Publication Date Title
JP6239857B2 (en) Imaging apparatus and control method thereof
US10545312B2 (en) Focus detection apparatus, control method thereof, and storage medium storing program
JP6249825B2 (en) Imaging device, control method thereof, and control program
JP6239855B2 (en) Focus adjustment apparatus, focus adjustment method and program, and imaging apparatus
JP6381266B2 (en) IMAGING DEVICE, CONTROL DEVICE, CONTROL METHOD, PROGRAM, AND STORAGE MEDIUM
JP2016038414A (en) Focus detection device, control method thereof, and imaging apparatus
JP6746359B2 (en) Image processing device, imaging device, image processing method, program, and storage medium
JP6700986B2 (en) Image processing device, imaging device, image processing method, and program
CN110312957B (en) Focus detection apparatus, focus detection method, and computer-readable storage medium
JP2016163220A (en) Image processing device, photographing device, image processing method, program, and recording medium
JP2017158018A (en) Image processing apparatus, control method of the same, and imaging apparatus
JP2015194736A (en) Imaging device and method for controlling the same
JP2015210285A (en) Imaging device, manufacturing method of the same, program thereof and recording medium
JP6254843B2 (en) Image processing apparatus and control method thereof
JP6285683B2 (en) Imaging apparatus and control method thereof
JP6862102B2 (en) Control device, image pickup device, control method, program, and storage medium
JP2015145970A (en) Imaging device, control method of the same, program abd recording medium
JP6525829B2 (en) Control device, imaging device, control method, program, and storage medium
JP7022575B2 (en) Focus detectors and methods, and imaging devices
JP2015225310A (en) Image capturing device, control method therefor, program, and storage medium
CN113596431B (en) Image processing apparatus, image capturing apparatus, image processing method, and storage medium
JP2015225311A (en) Image capturing device, control method therefor, program, and storage medium
JP6526294B2 (en) Imaging device and lens device
JP7005209B2 (en) Image pickup device and its control method
JP2018019348A (en) Imaging device, image processing method, image processing system, and image processing program

Legal Events

Date Code Title Description
A621 Written request for application examination

Free format text: JAPANESE INTERMEDIATE CODE: A621

Effective date: 20160510

A977 Report on retrieval

Free format text: JAPANESE INTERMEDIATE CODE: A971007

Effective date: 20170131

A131 Notification of reasons for refusal

Free format text: JAPANESE INTERMEDIATE CODE: A131

Effective date: 20170206

A521 Request for written amendment filed

Free format text: JAPANESE INTERMEDIATE CODE: A523

Effective date: 20170405

TRDD Decision of grant or rejection written
A01 Written decision to grant a patent or to grant a registration (utility model)

Free format text: JAPANESE INTERMEDIATE CODE: A01

Effective date: 20171006

A61 First payment of annual fees (during grant procedure)

Free format text: JAPANESE INTERMEDIATE CODE: A61

Effective date: 20171102

R151 Written notification of patent or utility model registration

Ref document number: 6239857

Country of ref document: JP

Free format text: JAPANESE INTERMEDIATE CODE: R151