WO2014133076A1 - 撮像素子および電子機器 - Google Patents
撮像素子および電子機器 Download PDFInfo
- Publication number
- WO2014133076A1 WO2014133076A1 PCT/JP2014/054865 JP2014054865W WO2014133076A1 WO 2014133076 A1 WO2014133076 A1 WO 2014133076A1 JP 2014054865 W JP2014054865 W JP 2014054865W WO 2014133076 A1 WO2014133076 A1 WO 2014133076A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- imaging
- area
- region
- unit
- image
- Prior art date
Links
- 238000003384 imaging method Methods 0.000 title claims abstract description 190
- 238000001514 detection method Methods 0.000 claims abstract description 40
- 230000003287 optical effect Effects 0.000 claims abstract description 31
- 238000004458 analytical method Methods 0.000 claims description 17
- 238000009826 distribution Methods 0.000 claims description 4
- 238000012545 processing Methods 0.000 description 57
- 238000000034 method Methods 0.000 description 28
- 230000008569 process Effects 0.000 description 28
- 230000000875 corresponding effect Effects 0.000 description 21
- 238000009825 accumulation Methods 0.000 description 14
- 238000010586 diagram Methods 0.000 description 11
- 238000012546 transfer Methods 0.000 description 11
- 238000009792 diffusion process Methods 0.000 description 6
- 239000004973 liquid crystal related substance Substances 0.000 description 6
- 238000012986 modification Methods 0.000 description 5
- 230000004048 modification Effects 0.000 description 5
- 230000003321 amplification Effects 0.000 description 4
- 230000008859 change Effects 0.000 description 4
- 238000003199 nucleic acid amplification method Methods 0.000 description 4
- 238000007781 pre-processing Methods 0.000 description 4
- 241000724291 Tobacco streak virus Species 0.000 description 3
- 230000002093 peripheral effect Effects 0.000 description 3
- 239000012071 phase Substances 0.000 description 3
- 230000004043 responsiveness Effects 0.000 description 3
- 241001465754 Metazoa Species 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 2
- 230000006835 compression Effects 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000005286 illumination Methods 0.000 description 2
- 238000002360 preparation method Methods 0.000 description 2
- 238000003860 storage Methods 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000003708 edge detection Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 230000004907 flux Effects 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 238000002844 melting Methods 0.000 description 1
- 230000008018 melting Effects 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 238000002161 passivation Methods 0.000 description 1
- 238000003825 pressing Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 229920006395 saturated elastomer Polymers 0.000 description 1
- 229910052710 silicon Inorganic materials 0.000 description 1
- 239000010703 silicon Substances 0.000 description 1
- 229910000679 solder Inorganic materials 0.000 description 1
- 239000007790 solid phase Substances 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/1464—Back illuminated imager structures
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/70—SSIS architectures; Circuits associated therewith
- H04N25/703—SSIS architectures incorporating pixels for producing signals other than image signals
- H04N25/704—Pixels specially adapted for focusing, e.g. phase difference pixel sets
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/1462—Coatings
- H01L27/14621—Colour filter arrangements
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/14641—Electronic components shared by two or more pixel-elements, e.g. one amplifier shared by two pixel elements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/67—Focus control based on electronic image sensor signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/67—Focus control based on electronic image sensor signals
- H04N23/672—Focus control based on electronic image sensor signals based on the phase difference signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/72—Combination of two or more compensation controls
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/10—Circuitry of solid-state image sensors [SSIS]; Control thereof for transforming different wavelengths into image signals
- H04N25/11—Arrangement of colour filter arrays [CFA]; Filter mosaics
- H04N25/13—Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements
- H04N25/134—Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements based on three different wavelength filter elements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/40—Extracting pixel data from image sensors by controlling scanning circuits, e.g. by modifying the number of pixels sampled or to be sampled
- H04N25/44—Extracting pixel data from image sensors by controlling scanning circuits, e.g. by modifying the number of pixels sampled or to be sampled by partially reading an SSIS array
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/40—Extracting pixel data from image sensors by controlling scanning circuits, e.g. by modifying the number of pixels sampled or to be sampled
- H04N25/44—Extracting pixel data from image sensors by controlling scanning circuits, e.g. by modifying the number of pixels sampled or to be sampled by partially reading an SSIS array
- H04N25/441—Extracting pixel data from image sensors by controlling scanning circuits, e.g. by modifying the number of pixels sampled or to be sampled by partially reading an SSIS array by reading contiguous pixels from selected rows or columns of the array, e.g. interlaced scanning
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/70—SSIS architectures; Circuits associated therewith
- H04N25/702—SSIS architectures characterised by non-identical, non-equidistant or non-planar pixel layout
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/70—SSIS architectures; Circuits associated therewith
- H04N25/703—SSIS architectures incorporating pixels for producing signals other than image signals
- H04N25/706—Pixels for exposure or ambient light measuring
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/70—SSIS architectures; Circuits associated therewith
- H04N25/79—Arrangements of circuitry being divided between different or multiple substrates, chips or circuit boards, e.g. stacked image sensors
Definitions
- the present invention relates to an image sensor and an electronic device.
- an electronic device including an imaging element (hereinafter referred to as a laminated imaging element) in which a backside illumination imaging chip and a signal processing chip are stacked (see Patent Document 1).
- the multilayer imaging element is laminated so that the back-illuminated imaging chip and the signal processing chip are connected to each other through micro bumps.
- the image pickup device picks up the light incident through the optical system under the first image pickup condition and generates a detection signal for performing focus detection of the optical system, and the optical element. And a second imaging region for generating an image signal by imaging light incident through the system under a second imaging condition different from the first imaging condition.
- the imaging device captures the light incident through the optical system under the first imaging condition and generates a detection signal for performing focus detection of the optical system, and the optical A second imaging region for imaging light incident through the system under a second imaging condition different from the first imaging condition and generating a signal used for exposure calculation.
- the imaging device captures the light incident through the optical system under the first imaging condition and generates a detection signal for performing focus detection of the optical system, and the optical element.
- the frame rate set by the first imaging condition is higher than the frame rate set by the second imaging condition. It is preferable.
- the area of the second imaging region is preferably larger than the area of the first imaging region.
- the electronic device sets the imaging element according to any one of the first to fifth aspects, the arrangement position of the first imaging area, and the arrangement position of the second imaging area. And comprising.
- the electronic device of the sixth aspect further includes an analysis unit that analyzes the luminance distribution of the subject, and the setting unit is configured based on the luminance distribution of the subject analyzed by the analysis unit. It is preferable to set the arrangement position of the first imaging area and the arrangement position of the second imaging area.
- the setting unit sets the first imaging region and the second imaging region in the region determined to include the main subject based on the analysis result of the analysis unit. It is preferable to do.
- the setting unit may set the two imaging regions in a region other than the region determined to include the main subject based on the analysis result of the analysis unit. preferable.
- the setting unit includes one imaging region and the second region in a region other than the region exceeding the predetermined luminance according to the analysis result of the analysis unit. It is preferable to set an imaging region.
- a recording unit that records an image signal
- an instruction unit that instructs the recording unit to record the image signal
- the setting unit preferably cancels the setting of the first imaging condition and the setting of the second imaging condition when the instruction unit instructs the recording unit to record the image signal.
- a plurality of main imaging pre-processing can be performed with high accuracy, and an electronic device that is easy to use can be realized.
- FIG. 1 is a cross-sectional view of the multilayer image sensor 100.
- the imaging device 100 includes a backside illumination imaging chip 113 that outputs a pixel signal corresponding to incident light, a signal processing chip 111 that processes the pixel signal, and a memory chip 112 that stores the pixel signal.
- the imaging chip 113, the signal processing chip 111, and the memory chip 112 are stacked, and are electrically connected to each other by a conductive bump 109 such as Cu.
- incident light is incident mainly in the positive direction of the Z axis indicated by the white arrow.
- the surface on the side where incident light is incident is referred to as a back surface.
- the left direction of the paper orthogonal to the Z axis is the X axis plus direction
- the front side of the paper orthogonal to the Z axis and the X axis is the Y axis plus direction.
- the coordinate axes are displayed so that the orientation of each figure can be understood with reference to the coordinate axes of FIG.
- the imaging chip 113 is a back-illuminated MOS image sensor.
- the PD layer 106 is disposed on the back side of the wiring layer 108.
- the PD layer 106 includes a plurality of PDs (photodiodes) 104 that are two-dimensionally arranged and store charges corresponding to incident light, and transistors 105 that are provided corresponding to the PDs 104.
- a color filter 102 is provided on the incident light incident side of the PD layer 106 via a passivation film 103.
- the color filter 102 has a plurality of types that transmit different wavelength regions, and has a specific arrangement corresponding to each of the PDs 104. The arrangement of the color filter 102 will be described later.
- a set of the color filter 102, the PD 104, and the transistor 105 forms one pixel.
- a microlens 101 is provided on the incident light incident side of the color filter 102 corresponding to each pixel.
- the microlens 101 condenses incident light toward the corresponding PD 104.
- the wiring layer 108 includes a wiring 107 that transmits a pixel signal from the PD layer 106 to the signal processing chip 111.
- the wiring 107 may be multilayer, and a passive element and an active element may be provided.
- a plurality of bumps 109 are arranged on the surface of the wiring layer 108.
- the plurality of bumps 109 are aligned with the plurality of bumps 109 provided on the opposing surfaces of the signal processing chip 111, and the imaging chip 113 and the signal processing chip 111 are pressed and aligned.
- the bumps 109 are joined and electrically connected.
- a plurality of bumps 109 are arranged on the mutually facing surfaces of the signal processing chip 111 and the memory chip 112.
- the bumps 109 are aligned with each other, and the signal processing chip 111 and the memory chip 112 are pressurized, so that the aligned bumps 109 are joined and electrically connected.
- the bonding between the bumps 109 is not limited to Cu bump bonding by solid phase diffusion, and micro bump bonding by solder melting may be employed. Further, for example, about one bump 109 may be provided for one unit region described later. Therefore, the size of the bump 109 may be larger than the pitch of the PD 104. Further, a bump larger than the bump 109 corresponding to the pixel region may be provided in a peripheral region other than the pixel region where the pixels are arranged.
- the signal processing chip 111 has TSVs (silicon through electrodes) 110 that connect circuits provided on the front and back surfaces to each other.
- the TSV 110 is preferably provided in the peripheral area.
- the TSV 110 may also be provided in the peripheral area of the imaging chip 113 and the memory chip 112.
- FIG. 2 is a diagram for explaining the pixel array and the unit area 131 of the imaging chip 113.
- a state where the imaging chip 113 is observed from the back side is shown.
- 20 million or more pixels are arranged in a matrix in the pixel region.
- 16 pixels of adjacent 4 pixels ⁇ 4 pixels form one unit region 131.
- the grid lines in the figure indicate the concept that adjacent pixels are grouped to form a unit region 131.
- the number of pixels forming the unit region 131 is not limited to this, and may be about 1000, for example, 32 pixels ⁇ 64 pixels, or more or less (for example, one pixel).
- the unit area 131 includes four so-called Bayer arrays, which are composed of four pixels of green pixels Gb, Gr, blue pixels B, and red pixels R, vertically and horizontally.
- the green pixel is a pixel having a green filter as the color filter 102, and receives light in the green wavelength band of incident light.
- a blue pixel is a pixel having a blue filter as the color filter 102 and receives light in the blue wavelength band
- a red pixel is a pixel having a red filter as the color filter 102 and receiving light in the red wavelength band. Receive light.
- a plurality of blocks are defined so as to include at least one unit region 131 per block, and each block can control pixels included in each block with different control parameters. That is, it is possible to acquire imaging signals having different imaging conditions between a pixel group included in a certain block and a pixel group included in another block.
- the control parameters are a frame rate, a gain, a thinning rate, the number of addition rows or addition columns to which pixel signals are added, the charge accumulation time or accumulation count, the number of digitization bits, and the like.
- the control parameter may be a parameter in image processing after obtaining an image signal from a pixel.
- FIG. 3 is a circuit diagram corresponding to the unit area 131 of the imaging chip 113.
- a rectangle surrounded by a dotted line typically represents a circuit corresponding to one pixel. Note that at least some of the transistors described below correspond to the transistor 105 in FIG.
- the unit region 131 is formed of 16 pixels.
- the 16 PDs 104 corresponding to the respective pixels are respectively connected to the transfer transistors 302, and the gates of the transfer transistors 302 are connected to the TX wiring 307 to which transfer pulses are supplied.
- the TX wiring 307 is commonly connected to the 16 transfer transistors 302.
- each transfer transistor 302 is connected to the source of the corresponding reset transistor 303, and a so-called floating diffusion FD between the drain of the transfer transistor 302 and the source of the reset transistor 303 is connected to the gate of the amplification transistor 304.
- the drain of the reset transistor 303 is connected to a Vdd wiring 310 to which a power supply voltage is supplied, and the gate thereof is connected to a reset wiring 306 to which a reset pulse is supplied.
- the reset wiring 306 is commonly connected to the 16 reset transistors 303.
- each amplification transistor 304 is connected to a Vdd wiring 310 to which a power supply voltage is supplied.
- the source of each amplification transistor 304 is connected to the drain of each corresponding transistor 305.
- Each gate of the selection transistor is connected to a decoder wiring 308 to which a selection pulse is supplied.
- the decoder wiring 308 is provided independently for each of the 16 selection transistors 305.
- the source of each selection transistor 305 is connected to a common output wiring 309.
- the load current source 311 supplies current to the output wiring 309. That is, the output wiring 309 for the selection transistor 305 is formed by a source follower. Note that the load current source 311 may be provided on the imaging chip 113 side or may be provided on the signal processing chip 111 side.
- the PD 104 converts the incident light received into charges and accumulates them. Thereafter, when the transfer pulse is applied again without the reset pulse being applied, the accumulated charge is transferred to the floating diffusion FD, and the potential of the floating diffusion FD changes from the reset potential to the signal potential after the charge accumulation. .
- a selection pulse is applied to the selection transistor 305 through the decoder wiring 308, a change in the signal potential of the floating diffusion FD is transmitted to the output wiring 309 through the amplification transistor 304 and the selection transistor 305. Thereby, a pixel signal corresponding to the reset potential and the signal potential is output from the unit pixel to the output wiring 309.
- the reset wiring 306 and the TX wiring 307 are common to the 16 pixels forming the unit region 131. That is, the reset pulse and the transfer pulse are simultaneously applied to all 16 pixels. Accordingly, all the pixels forming the unit region 131 start charge accumulation at the same timing and end charge accumulation at the same timing. However, the pixel signal corresponding to the accumulated charge is selectively output from the output wiring 309 by sequentially applying the selection pulse to each selection transistor 305. Further, the reset wiring 306, the TX wiring 307, and the output wiring 309 are provided separately for each unit region 131.
- the charge accumulation time can be controlled for each unit region 131.
- pixel signals having different frame rates can be output between the unit regions 131.
- the other unit region 131 is caused to repeatedly accumulate charges several times and each time a pixel signal is output, so that Each frame for moving images can be output at a different frame rate between the unit areas 131.
- FIG. 4 is a block diagram showing a functional configuration of the image sensor 100.
- the analog multiplexer 411 sequentially selects the 16 PDs 104 forming the unit area 131 and outputs each pixel signal to the output wiring 309 provided corresponding to the unit area 131.
- the multiplexer 411 is formed on the imaging chip 113 together with the PD 104.
- the pixel signal output via the multiplexer 411 is supplied to the signal processing chip 111 by a signal processing circuit 412 that performs correlated double sampling (CDS) / analog / digital (A / D) conversion. D conversion is performed.
- CDS correlated double sampling
- a / D converted pixel signal is transferred to the demultiplexer 413 and stored in the pixel memory 414 corresponding to each pixel.
- the demultiplexer 413 and the pixel memory 414 are formed in the memory chip 112.
- the arithmetic circuit 415 processes the pixel signal stored in the pixel memory 414 and passes it to the subsequent image processing unit.
- the arithmetic circuit 415 may be provided in the signal processing chip 111 or may be provided in the memory chip 112. Note that FIG. 4 shows connections for one unit region 131, but actually these exist for each unit region 131 and operate in parallel. However, the arithmetic circuit 415 does not have to exist for each unit region 131. For example, one arithmetic circuit 415 may perform sequential processing while sequentially referring to the values of the pixel memory 414 corresponding to each unit region 131. Good.
- the output wiring 309 is provided corresponding to each of the unit areas 131. Since the image pickup device 100 has the image pickup chip 113, the signal processing chip 111, and the memory chip 112 laminated, by using electrical connection between the chips using the bump 109 for the output wiring 309, each chip is arranged in the surface direction. Wiring can be routed without increasing the size.
- FIG. 5 is a block diagram illustrating the configuration of the image pickup apparatus 1 having the image pickup device 100 described above.
- the imaging apparatus 1 includes an imaging optical system 10, an imaging unit 20, an image processing unit 30, a work memory 40, a display unit 50, a recording unit 60, and a control unit 70.
- the imaging optical system 10 is composed of a plurality of lenses, and guides the light flux from the object scene to the imaging unit 20.
- the imaging optical system 10 may be configured integrally with the imaging device 1 or may be configured to be replaceable with respect to the imaging device 1. Further, the imaging optical system 10 may include a focus lens or a zoom lens.
- the imaging unit 20 includes the above-described imaging device 100 and a driving unit 21 that drives the imaging device 100.
- the image sensor 100 is driven and controlled by a control signal output from the drive unit 21, so that independent storage control can be performed in units of blocks as described above.
- the controller 70 gives instructions to the drive unit 21 such as the position, shape, and range of the block.
- the image processing unit 30 performs image processing on the image data captured by the imaging unit 20 in cooperation with the work memory 40.
- the image processing unit 30 performs detection processing of a main subject included in an image in addition to normal image processing (color signal processing, gamma correction, etc.). Detection of the main subject by the image processing unit 30 can be performed using a known face detection function. In addition to face detection, a human body included in an image may be detected as a main subject as described in, for example, Japanese Unexamined Patent Application Publication No. 2010-16621 (US2010 / 0002940).
- the display unit 50 includes, for example, a liquid crystal display panel 51, and displays images (still images and moving images) and various information captured by the imaging unit 20, and displays an operation input screen.
- the display unit 50 has a configuration in which a touch panel 52 is laminated on the display surface of the liquid crystal display panel 51.
- the touch panel 52 outputs a signal indicating a position where the user touches the liquid crystal display panel 51.
- the recording unit 60 stores various data such as image data in a storage medium such as a memory card.
- the control unit 70 has a CPU and controls the overall operation of the imaging apparatus 1.
- the control unit 70 divides the imaging surface of the imaging device 100 (imaging chip 113) into a plurality of blocks, and acquires images at different frame rates and gains between the blocks.
- the control unit 70 instructs the drive unit 21 of the block position, shape, range, and control parameters for each block.
- control unit 70 calculates the focus adjustment state by the imaging optical system 10 by the AF calculation unit 71.
- the control unit 70 further performs exposure calculation by the AE and AWB calculation unit 72 so that proper exposure can be obtained.
- FIG. 6 is a diagram illustrating the arrangement of the AF area and the AE area in the imaging device 100 (imaging chip 113).
- the hatched portion represents the AE area
- the white portion represents the AF area.
- the AF area and the AE area are determined in advance so as to be arranged in a checkered pattern. For example, when the live view image is acquired, the control unit 70 performs focus detection processing by the AF calculation unit 71 using a signal output from the AF area of the image sensor 100.
- the live view image is also referred to as a preview image before actual imaging, and refers to a monitor image acquired by the imaging device 100 at a predetermined frame rate (for example, 30 fps).
- the focus detection process is performed by, for example, a contrast detection method. Specifically, the position of the focus lens of the imaging optical system 10 is adjusted so as to increase the contrast of an image composed of signals output from the AF area while moving the position of the focus lens of the imaging optical system 10.
- the focus detection process may be performed by a phase difference detection method.
- pixels for focus detection are provided in the AF area of the image sensor 100 (imaging chip 113).
- a phase difference detection calculation is performed using an output signal from a focus detection pixel, thereby detecting a focus adjustment state (specifically, a defocus amount) by the imaging optical system 10.
- the focus detection pixel and the phase difference detection calculation are well known as described in, for example, Japanese Patent Application Laid-Open No. 2009-94881, and thus detailed description thereof is omitted.
- control unit 70 performs an exposure calculation process in the AE / AWB calculation unit 72 using a signal output from the AE area when the live view image is acquired. For example, the AE / AWB calculation unit 72 determines exposure (frame rate, gain, etc.) so that the average level of the signal output from the AE region approaches a predetermined level. The controller 70 further determines a white balance adjustment value based on the signal output from the AE area. Furthermore, the control unit 70 generates an image for monitoring based on a signal output from the AE area, and displays it on the display unit 50 as the live view image.
- the control unit 70 sends an instruction to the drive unit 21 so as to perform accumulation control of the image sensor 100 (image pickup chip 113) separately in the AF area and the AE area.
- the accumulation control is performed separately for the AF area and the AE area before a real imaging (still image recording or moving image recording) instruction is given by a release switch (not shown).
- the control parameters that are the pre-imaging processing for the AF area are, for example, a frame rate of 120 fps and a gain that is high in the AE area.
- the control parameters, which are the main imaging pre-processing for the AE area are, for example, a frame rate of 30 fps, which is slower than the AF area, and a gain lower than the gain of the AF area.
- the arrangement of the AF area and the AE area in the image sensor 100 is not limited to the arrangement illustrated in FIG. 6, and the sizes of the AF area and the AE area may be changed as appropriate. As illustrated in FIG. 7, it may be changed smaller than the case of FIG. 6, or may be changed greatly compared to the case of FIG. 6.
- FIG. 8 is a diagram exemplifying a case where the area of the AE region is made wider than the area of the AF region and the area ratio of both is changed.
- the live view image described above is generated by thinning out a part of the pixels of the image sensor 100, for example, an AF area is set corresponding to the thinned out pixels, and the live view image corresponds to the pixels of the live view image. If the AE area is set, the quality of the live view image will not deteriorate.
- the AF area is to be arranged only at a predetermined position such as the center of the screen, the AF area may be arranged at a predetermined size and only at a predetermined position on the screen.
- FIG. 9 is a flowchart for explaining the flow of the photographing operation executed by the control unit 70 of the imaging device 1 in the first embodiment.
- the control unit 70 repeatedly activates the processing of FIG. 9 when an ON-OFF switch (not shown) is turned on and power is supplied to each unit of the imaging apparatus 1.
- the control unit 70 determines control parameters such as a frame rate and a gain for the AF area and the AE area, and proceeds to step S102. For example, values to be applied in steps S102, S104, and S105, which will be described later, are read out from the program data and prepared.
- step S102 the control unit 70 sends an instruction to the drive unit 21 to start imaging by the imaging unit 20.
- the acquisition of the image starting in step S102 is performed by setting control parameters for the AE region, for example, with the substantially entire area of the imaging surface of the image sensor 100 as the AE region.
- the control unit 70 causes the image processing unit 30 to perform image processing on live view image data based on a signal output from the imaging unit 20 and then causes the display unit 50 to display the live view image data.
- step S103 the control unit 70 determines whether or not the release half-press operation has been performed.
- the release half-press operation is used as an instruction to start the main imaging pre-processing (imaging preparation) for the imaging apparatus 1.
- a release button (not shown) is half-pressed, the control unit 70 makes a positive determination in step S103 and proceeds to step S104. If the half-press operation is not performed, the control unit 70 makes a negative determination in step S103 and Repeat the determination process. When the determination process is repeated, it waits for a release half-press operation.
- step S ⁇ b> 103 may be omitted when the tap operation on the release icon displayed on the liquid crystal display panel 51 is detected without determining the image pickup instruction without providing the image pickup apparatus 1 with the release button. .
- the control unit 70 performs processing of steps S104 and S105, which will be described later, from the start of imaging in step S102.
- Different control parameters (frame rate, gain) are applied to the AF area and AE area.
- the control unit 70 may perform the same processing as the tap operation on the release icon.
- the actual imaging pre-processing (imaging preparation) may be performed within a few seconds after is turned on.
- step S104 the control unit 70 sends an instruction to the driving unit 21 to apply control parameters such as a frame rate and gain for the AF area to the AF area of the image sensor 100.
- step S ⁇ b> 105 the control unit 70 sends an instruction to the drive unit 21 to apply control parameters such as the frame rate and gain for the AE region to the AE region of the image sensor 100. Accordingly, the control unit 70 performs fine adjustment of the control parameter, determination of the white balance adjustment value, and display of the live view image based on the signal output from the AE area at the frame rate.
- step S106 the control unit 70 starts AF processing. Specifically, the AF calculation unit 71 starts focus detection processing based on the signal output from the AF area, and proceeds to step S107. Thereby, the focus adjustment of the imaging optical system 10 can be performed based on the signal output from the AF area at the frame rate.
- step S107 the control unit 70 determines whether or not the AF process is finished. For example, when the contrast of the image obtained from the AF area is equal to or greater than a predetermined value, the control unit 70 makes a positive determination in step S107, and when the contrast of the image obtained from the AF area is less than the predetermined value, the control unit 70 makes a negative determination. To do. When a negative determination is made, the determination process is repeated while continuing the focus detection process.
- step S108 the control unit 70 causes the AE and AWB calculation unit 72 to calculate the exposure condition and white balance adjustment value to be applied during shooting based on the signal output from the AE area after focus adjustment, and proceeds to step S109.
- step S109 the control unit 70 determines whether or not the release full-press operation has been performed.
- the release full pressing operation is used as a main imaging instruction for the imaging apparatus 1.
- the control unit 70 makes a positive determination in step S109 and proceeds to step S110. If the release button is not fully pressed, the control unit 70 makes a negative determination in step S109. Return to S108.
- the control unit 70 releases the imaging of the imaging unit 20 by the control parameter of the AF area and the control parameter of the AE area when a release button (not shown) is fully pressed.
- step S109 is affirmed when the displayed release icon is tapped. What is necessary is just to judge.
- step S110 the control unit 70 sends an instruction to the driving unit 21, sets control parameters (exposure time, gain, etc.) necessary for the exposure conditions for photographing (main imaging) calculated in step S108, and proceeds to step S111. .
- step S111 the control unit 70 executes a photographing process, stores the acquired image data in a memory card or the like by the recording unit 60, and ends the process of FIG.
- the control unit 70 may change the imaging condition of the pixel corresponding to the high-luminance subject when there is a high-luminance subject in part of the object scene. Then, based on the images of a plurality of frames acquired during a predetermined time before and after the release full-press operation, slow reproduction moving image data is generated and recorded.
- Slow playback moving image data refers to moving image data that is played back at a frame rate (for example, 15 fps) slower than the frame rate (for example, 30 fps) obtained by the image sensor 100.
- the control unit 70 generates slow playback moving image data as follows. That is, temporarily stored in the work memory 40 for displaying the above-described live view image from a first predetermined time (for example, 0.6 seconds before) to time t1 from the release full press operation time (t1). A plurality of frame images based on the AE area (the frame rate is set to the frame rate set in the AE area.
- 0.6 seconds when acquired at 30 fps is 18 frames
- a plurality of frame images stored in the work memory 40 from the time t1 until the second predetermined time (for example, after 0.4 seconds) (the frame rate is the same as the frame rate set in the AE area before the main imaging instruction)
- 0.4 seconds when acquired at 30 fps is 12 frames.
- playback is performed based on a plurality of frame images (a total of 30 images) stored in the work memory 40 for one second (0.6 seconds before time t1 to 0.4 seconds after time t1) with time t1 interposed therebetween.
- Slow playback moving image data having a time of about 2 seconds is generated. In this way, slow playback moving image data based on frame images acquired before and after the release full-press operation is obtained.
- the slow playback moving image data is generated by the image processing unit 30 as MPEG data or JPEG data.
- the imaging apparatus 1 can obtain signals used for exposure calculation processing and signals used for detection of the focus adjustment state at appropriate levels, exposure calculation and focus adjustment can be performed quickly and accurately. As a result, a user-friendly electronic device can be realized.
- the control unit 70 releases the imaging of the imaging unit 20 by the control parameter of the AF area and the control parameter of the AE area in response to a full press of a release switch (not shown), and performs imaging under imaging conditions suitable for main imaging. Since the unit 20 is driven, a better still image / moving image can be obtained.
- the control unit 70 can improve the responsiveness of the focus detection process by acquiring a signal used for the focus detection process in the second area by applying a higher frame rate than the first area.
- control unit 70 Since the control unit 70 performs the exposure calculation process with the gain setting of the first area set to a lower gain setting than the second area, the calculation can be performed with high accuracy while suppressing the influence of noise.
- control unit 70 determines the first area and the second area in advance, the exposure is determined based on a predetermined position in the screen, or the focus is adjusted based on the predetermined position in the screen. Can be.
- FIG. 10 is a diagram for explaining the relationship between the AF area and the AE area and the live view image in the second embodiment.
- a person is included in the center of the screen of the live view image
- a tree is included in the upper left portion of the screen of the live view image
- the sun is included in the upper right portion of the screen of the live view image.
- the control unit 70 sends an instruction to the image processing unit 30 to perform a known scene recognition process on the live view image data.
- the image processing unit 30 extracts a high luminance area A, a main subject area B, and a background area C from the live view image by performing a scene recognition process.
- the image processing unit 30 sets, for example, a range in which the signal level of the live view image data exceeds a predetermined high luminance threshold as the high luminance area A.
- the image processing unit 30 sets the range including the human body detected as described above as the main subject region B. Note that not only a person but also an animal such as a pet may be detected, and a range including this animal may be set as the main subject region B.
- the image processing unit 30 performs edge detection at a position excluding the high luminance area A and the main subject area B, and sets a range where the density of detected edges exceeds a predetermined value as the background area C. If there is no data exceeding the high brightness threshold in the live view image data, it is assumed that there is no high brightness area A.
- the control unit 70 arranges the AF area and the AE area, for example, in a checkered pattern in the main subject area B.
- the hatched portion in the main subject region B represents the AE region
- the white portion represents the AF region.
- the arrangement of the AF area and the AE area in the main subject area B is not limited to the arrangement illustrated in FIG. 10, and the sizes of the AF area and the AE area may be appropriately changed.
- the area ratio may be changed as appropriate so that the area of the AE area is larger than the area of the AF area.
- the AF area when it is desired to arrange the AF area only at a predetermined position such as the central portion in the main subject area B, the AF area may be arranged at a predetermined size and only at the predetermined position. Further, the control unit 70 may arrange the AF area and the AE area so that a part of the AF area and the AE area protrudes from the main subject area B.
- control unit 70 sets the background area C as the AE area.
- the hatched portion in the background area C represents the AE area.
- control unit 70 does not determine an AF area and an AE area for a portion exceeding the high brightness threshold (high brightness area A in the example of FIG. 10).
- high brightness area A in the example of FIG. 10
- a portion with a vertical line in the high luminance area A indicates that it is neither an AF area nor an AE area.
- the control unit 70 includes a portion (a white background portion in FIG. 10) that does not belong to any of the high luminance area A, the main subject area B, and the background area C on the screen of the live view image. AE area.
- FIG. 11 is a diagram illustrating three settings.
- Setting 1 is a control parameter (frame) suitable for high brightness so that the output signal does not saturate, for example, by increasing the frame rate and decreasing the gain for the high brightness area A (that is, neither AF area nor AE area). Rate, gain, etc.).
- Setting 2 is to set control parameters (frame rate, gain, etc.) suitable for focus detection processing for the AF area in the main subject area B, and to set the AE in the main subject area B as in the first embodiment. It shows that control parameters (frame rate, gain, etc.) suitable for exposure calculation processing are set for the area, as in the first embodiment.
- Setting 3 is the same as the AE area in the main subject area B with respect to the background area C and the portion that does not belong to the high brightness area A or the main subject area B (the white background portion in FIG. 10). It shows that control parameters (frame rate, gain, etc.) suitable for the above are set.
- the control unit 70 sends an instruction to the driving unit 21 to cause the imaging device 100 (imaging chip 113) to perform the above three settings.
- the control parameters frame rate, gain, etc.
- the accumulation control is performed, for example, after the scene recognition by the image processing unit 30 until the actual imaging instruction is performed.
- the focus detection process is performed based on the image acquired from the AF area
- the exposure calculation and the white balance calculation are performed based on the image acquired from the AE area
- the image acquired in the AE area and the AF area are also calculated.
- the live view image is displayed based on the images acquired in the area that is not the AE area.
- the control parameters for the AF area are, for example, a frame rate of 120 fps and a gain that is higher in the AE area.
- the control parameters for the AE area are, for example, a frame rate of 30 fps, which is slower than the AF area, and a gain lower than the gain of the AF area.
- the reason why the image used for the focus detection process is acquired in the AF area by applying a higher frame rate than the AE area is to improve the responsiveness of the focus detection process. Further, the reason why the gain used in the exposure calculation process and the determination of the white balance adjustment value is obtained by setting the gain in the AE area to be lower than the gain in the AF area is to perform an accurate calculation while avoiding the influence of noise.
- the control parameters for an area that is neither an AF area nor an AE area are, for example, a frame rate of 60 fps and a gain lower than that of the AE area. As a result, a high-luminance subject image that causes so-called whiteout can be acquired without whiteout.
- FIG. 12 is a flowchart for describing the flow of a photographing operation performed by the control unit 70 of the imaging device 1 in the second embodiment. Compared with the flowchart in the first embodiment illustrated in FIG. 9, the processes in steps S104B and 105B are different, and the difference will be mainly described.
- the acquisition of the image started in step S102 of FIG. 12 is performed by setting the control parameters for the AE region with the substantially entire area of the imaging surface of the image sensor 100 as the AE region, as in the case of FIG.
- the control unit 70 sets the live view image data based on the signal output from the imaging unit 20 with the above setting as a scene recognition target.
- the control unit 70 sends an instruction to the image processing unit 30 to perform a known scene recognition process on the live view image data.
- the image processing unit 30 performs the scene recognition process to divide into the high luminance area A, the main subject area B, and the background area C as described above.
- step S105B the control unit 70 sends an instruction to the driving unit 21 to perform the above-described three settings for each divided region.
- different images are obtained by applying different control parameters to the AF area, the AE area, and the area that is neither the AF area nor the AE area.
- the control unit 70 can finely adjust the control parameter and determine the white balance adjustment value based on the image acquired from the AE area. Further, a live view image can be displayed based on an image acquired in the AE area and an image acquired in an area that is neither an AF area nor an AE area.
- the imaging apparatus 1 can change the position to be subjected to exposure calculation and the position to be focused on the screen based on the scene analysis result.
- control unit 70 determines the first region and the second region within the range determined to include the main subject by the image processing unit 30 in the image, the target unit is targeted on the screen.
- the exposure can be determined and focus adjustment can be performed on the main subject.
- control unit 70 determines the first region outside the range determined by the image processing unit 30 in the image, the background other than the main subject can be included in the exposure calculation target.
- control unit 70 sets the first region and the second region in a region different from the region exceeding the predetermined luminance in the image, for example, when the signal from the imaging unit 20 is saturated These can be excluded from exposure calculation targets and focus adjustment targets.
- Imaging device 1 which concerns on 1st embodiment mentioned above and 2nd embodiment with a highly functional mobile telephone or a tablet terminal.
- a camera unit mounted on a high-function mobile phone (or tablet terminal) is configured using the multilayer image sensor 100.
- the control unit 70 performs a focus detection process based on the image acquired from the AF area.
- the control unit 70 determines the white balance adjustment value based on the image acquired from the AWB area.
- the fine adjustment of the control parameter is performed based on the image acquired from the AE area.
- the live view image is displayed based on an image acquired in the AE area and an image acquired in an area that is neither the AF area nor the AE area.
Landscapes
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Physics & Mathematics (AREA)
- Power Engineering (AREA)
- Condensed Matter Physics & Semiconductors (AREA)
- Computer Hardware Design (AREA)
- Microelectronics & Electronic Packaging (AREA)
- General Physics & Mathematics (AREA)
- Electromagnetism (AREA)
- Spectroscopy & Molecular Physics (AREA)
- Studio Devices (AREA)
- Transforming Light Signals Into Electric Signals (AREA)
- Automatic Focus Adjustment (AREA)
- Solid State Image Pick-Up Elements (AREA)
Abstract
Description
本発明の第2の態様によると、撮像素子は、光学系を介して入射した光を第1撮像条件で撮像して光学系の焦点検出を行う検出信号を生成する第1撮像領域と、光学系を介して入射した光を第1撮像条件とは異なる第2撮像条件で撮像して露出演算に用いる信号を生成する第2撮像領域と、を備える。
本発明の第3の態様によると、撮像素子は、光学系を介して入射した光を第1撮像条件で撮像して光学系の焦点検出を行う検出信号を生成する第1撮像領域と、光学系を介して入射した光を第1撮像条件とは異なる第2撮像条件で撮像してホワイトバランス調整に用いる信号を生成する第2撮像領域と、を備える。
本発明の第4の態様によると、第1~第3のいずれか一態様の撮像素子において、第1撮像条件により設定されたフレームレートは、第2撮像条件により設定されたフレームレートよりも高いことが好ましい。
本発明の第5の態様によると、第1~第4のいずれか一態様の撮像素子において、第2撮像領域の面積は、第1撮像領域の面積よりも大きいことが好ましい。
本発明の第6の態様によると、電子機器は、第1~第5のいずれか一態様の撮像素子と、第1撮像領域の配置位置と第2撮像領域の配置位置とを設定する設定部と、を備える。
本発明の第7の態様によると、第6の態様の電子機器において、被写体の輝度分布を解析する解析部を更に備え、設定部は、解析部により解析された被写体の輝度分布に基づいて第1撮像領域の配置位置と第2撮像領域の配置位置とを設定することが好ましい。
本発明の第8の態様によると、第7の態様の電子機器において、設定部は、解析部の解析結果により主要被写体を含むと判断された領域に第1撮像領域及び第2撮像領域を設定することが好ましい。
本発明の第9の態様によると、第8の態様の電子機器において、設定部は、解析部の解析結果により主要被写体を含むと判断された領域以外の領域に2撮像領域を設定することが好ましい。
本発明の第10の態様によると、第6~第9のいずれか一態様の電子機器において、設定部は、解析部の解析結果により所定輝度を超える領域以外の領域に1撮像領域及び第2撮像領域を設定することが好ましい。
本発明の第11の態様によると、第6~第10のいずれか一態様の電子機器において、画像信号を記録する記録部と、記録部に画像信号を記録するように指示する指示部と、を更に備え、設定部は、指示部により記録部に画像信号を記録するように指示された場合、第1撮像条件の設定と第2撮像条件の設定とを解除することが好ましい。
(第一の実施形態)
<積層型撮像素子の説明>
始めに、本発明の第一の実施形態による電子機器(例えば撮像装置1)に搭載する積層型撮像素子100について説明する。なお、この積層型撮像素子100は、本願出願人が先に出願した特願2012-139026号に記載されているものである。図1は、積層型撮像素子100の断面図である。撮像素子100は、入射光に対応した画素信号を出力する裏面照射型撮像チップ113と、画素信号を処理する信号処理チップ111と、画素信号を記憶するメモリチップ112とを備える。これら撮像チップ113、信号処理チップ111およびメモリチップ112は積層されており、Cu等の導電性を有するバンプ109により互いに電気的に接続される。
図5は、上述した撮像素子100を有する撮像装置1の構成を例示するブロック図である。図5において、撮像装置1は、撮像光学系10、撮像部20、画像処理部30、ワークメモリ40、表示部50、記録部60、および制御部70を有する。
第一の実施形態では、画面内にAF領域とAE領域という概念を導入し、上記複数のブロックに対応させる。図6は、撮像素子100(撮像チップ113)におけるAF領域およびAE領域の配置を例示する図である。図6において、斜線を付した部分がAE領域を表し、白い部分がAF領域を表す。第一の実施形態では、AF領域およびAE領域を市松模様状に配置するようにあらかじめ定めておく。制御部70は、例えばライブビュー画像取得時において、撮像素子100のAF領域から出力される信号を用いて、AF演算部71で焦点検出処理を行う。
図9は、第一の実施形態において、撮像装置1の制御部70が実行する撮影動作の流れを説明するフローチャートである。制御部70は、不図示のON-OFFスイッチがオン操作され、撮像装置1の各部に対して通電が行われている場合に、図9による処理を繰り返し起動させる。図9のステップS101において、制御部70は、AF領域およびAE領域用のフレームレート、ゲインなどの制御パラメータをそれぞれ決定してステップS102へ進む。例えば、後述するステップS102、S104、S105において適用する値を、プログラムデータから読み出して用意しておく。
(1)撮像装置1は、露出演算処理に用いる信号と焦点調節状態の検出に用いる信号とをそれぞれ適切なレベルで得られるから、露出演算や焦点調節をすみやかに精度よく行ない得る。この結果、使い勝手のよい電子機器を実現できる。
上述した第一の実施形態では、撮像素子100(撮像チップ113)におけるAF領域とAE領域の配置をあらかじめ定めておく例を説明したが、第二の実施形態では、撮像素子100(撮像チップ113)におけるAF領域とAE領域の配置を、ライブビュー画像に基づくシーン認識を経て決定する。なお、撮像装置1の構成は第一の実施形態において説明した図5の構成と同様であるので、構成についての説明は省略する。
図12は、第二の実施形態において、撮像装置1の制御部70が実行する撮影動作の流れを説明するフローチャートである。図9に例示した第一の実施形態におけるフローチャートと比べて、ステップS104Bおよび105Bの処理が異なるので、これらの相違点を中心に説明する。
(1)撮像装置1は、画面の中で露出演算の対象とする位置、および焦点調節の対象とする位置を、シーン解析結果に基づいて可変することができる。
上述した第一の実施形態および第二の実施形態に係る撮像装置1を、高機能携帯電話機、またはタブレット端末によって構成してもよい。この場合、高機能携帯電話機(またはタブレット端末)に搭載されるカメラユニットを、上記積層型撮像素子100を用いて構成する。
上述した説明では、AE領域から取得された画像に基づいて、制御パラメータの微調整と、ホワイトバランス調整値の決定とを行う例を説明した。この代わりに、AF領域およびAE領域と別にAWB領域を新たに設けてもよい。そして、撮像素子100(撮像チップ113)を、ライブビュー画像を取得する際、AF領域と、AE領域と、AWB領域とにおいてそれぞれ異なる制御パラメータ(フレームレート、ゲインなど)で蓄積制御する。
10…撮像光学系
20…撮像部
30…画像処理部
40…ワークメモリ
50…表示部
51…液晶表示パネル
52…タッチパネル
60…記録部
70…制御部
71…AF演算部
72…AE、AWB演算部
100…撮像素子
109…バンプ
111…信号処理チップ
112…メモリチップ
113…撮像チップ
131…単位領域
日本国特許出願2013年第037617号(2013年2月27日出願)
Claims (11)
- 光学系を介して入射した光を第1撮像条件で撮像して前記光学系の焦点検出を行う検出信号を生成する第1撮像領域と、
前記光学系を介して入射した光を前記第1撮像条件とは異なる第2撮像条件で撮像して画像信号を生成する第2撮像領域と、
を備える撮像素子。 - 光学系を介して入射した光を第1撮像条件で撮像して前記光学系の焦点検出を行う検出信号を生成する第1撮像領域と、
前記光学系を介して入射した光を前記第1撮像条件とは異なる第2撮像条件で撮像して露出演算に用いる信号を生成する第2撮像領域と、
を備える撮像素子。 - 光学系を介して入射した光を第1撮像条件で撮像して前記光学系の焦点検出を行う検出信号を生成する第1撮像領域と、
前記光学系を介して入射した光を前記第1撮像条件とは異なる第2撮像条件で撮像してホワイトバランス調整に用いる信号を生成する第2撮像領域と、
を備える撮像素子。 - 請求項1から請求項3のいずれか一項に記載の撮像素子において、
前記第1撮像条件により設定されたフレームレートは、前記第2撮像条件により設定されたフレームレートよりも高い撮像素子。 - 請求項1から請求項4のいずれか一項に記載の撮像素子において、
前記第2撮像領域の面積は、前記第1撮像領域の面積よりも大きい撮像素子。 - 請求項1から請求項5のいずれか一項に記載の撮像素子と、
前記第1撮像領域の配置位置と前記第2撮像領域の配置位置とを設定する設定部と、
を備える電子機器。 - 請求項6に記載の電子機器において、
被写体の輝度分布を解析する解析部を更に備え、
前記設定部は、前記解析部により解析された被写体の輝度分布に基づいて前記第1撮像領域の配置位置と前記第2撮像領域の配置位置とを設定する電子機器。 - 請求項7に記載の電子機器において、
前記設定部は、前記解析部の解析結果により主要被写体を含むと判断された領域に前記第1撮像領域及び前記第2撮像領域を設定する電子機器。 - 請求項8に記載の電子機器において、
前記設定部は、前記解析部の解析結果により主要被写体を含むと判断された領域以外の領域に前記2撮像領域を設定する電子機器。 - 請求項6から請求項9のいずれか一項に記載の電子機器において、
前記設定部は、前記解析部の解析結果により所定輝度を超える領域以外の領域に前記1撮像領域及び前記第2撮像領域を設定する電子機器。 - 請求項6から請求項10のいずれか一項に記載の電子機器において、
画像信号を記録する記録部と、
前記記録部に画像信号を記録するように指示する指示部と、を更に備え、
前記設定部は、前記指示部により前記記録部に画像信号を記録するように指示された場合、前記第1撮像条件の設定と前記第2撮像条件の設定とを解除する電子機器。
Priority Applications (7)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811159203.XA CN109348148B (zh) | 2013-02-27 | 2014-02-27 | 成像元件及电子设备 |
JP2015503018A JP6090425B2 (ja) | 2013-02-27 | 2014-02-27 | 撮像素子および電子機器 |
CN201480010575.XA CN105103537B (zh) | 2013-02-27 | 2014-02-27 | 成像元件及电子设备 |
EP14756999.0A EP2963917B1 (en) | 2013-02-27 | 2014-02-27 | Electronic apparatus comprising an image sensor |
US14/771,123 US10924697B2 (en) | 2013-02-27 | 2014-02-27 | Image sensor and electronic device having imaging regions for focus detection extending in first and second different directions |
US17/146,870 US11595604B2 (en) | 2013-02-27 | 2021-01-12 | Image sensor and imaging device including a plurality of semiconductor substrates |
US18/104,936 US20230199343A1 (en) | 2013-02-27 | 2023-02-02 | Image sensor and imaging device including a plurality of semiconductor substrates |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2013-037617 | 2013-02-27 | ||
JP2013037617 | 2013-02-27 |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/771,123 A-371-Of-International US10924697B2 (en) | 2013-02-27 | 2014-02-27 | Image sensor and electronic device having imaging regions for focus detection extending in first and second different directions |
US17/146,870 Continuation US11595604B2 (en) | 2013-02-27 | 2021-01-12 | Image sensor and imaging device including a plurality of semiconductor substrates |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2014133076A1 true WO2014133076A1 (ja) | 2014-09-04 |
Family
ID=51428335
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2014/054865 WO2014133076A1 (ja) | 2013-02-27 | 2014-02-27 | 撮像素子および電子機器 |
Country Status (5)
Country | Link |
---|---|
US (3) | US10924697B2 (ja) |
EP (1) | EP2963917B1 (ja) |
JP (5) | JP6090425B2 (ja) |
CN (2) | CN109348148B (ja) |
WO (1) | WO2014133076A1 (ja) |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2016052640A1 (ja) * | 2014-09-30 | 2016-04-07 | 株式会社ニコン | 電子機器、撮像装置、および、撮像システム |
CN107005635A (zh) * | 2014-09-30 | 2017-08-01 | 株式会社尼康 | 电子设备、再现装置、再现方法、记录介质及记录方法 |
CN107005633A (zh) * | 2014-09-30 | 2017-08-01 | 株式会社尼康 | 电子设备、再现装置、再现方法、记录介质及记录方法 |
CN107005636A (zh) * | 2014-09-30 | 2017-08-01 | 株式会社尼康 | 电子设备 |
JP2017225084A (ja) * | 2016-06-17 | 2017-12-21 | 株式会社ニコン | 撮像素子、電子機器、撮像方法、およびプログラム |
EP3203720A4 (en) * | 2014-09-30 | 2018-03-21 | Nikon Corporation | Electronic apparatus, reproduction device, reproduction method, recording medium, and recording method |
EP3203722A4 (en) * | 2014-09-30 | 2018-06-06 | Nikon Corporation | Electronic apparatus and image processing method |
JP2020036330A (ja) * | 2013-02-27 | 2020-03-05 | 株式会社ニコン | 撮像装置および撮像素子 |
JP2020101969A (ja) * | 2018-12-21 | 2020-07-02 | 株式会社モルフォ | 画像処理装置、画像処理方法およびプログラム |
CN112653821A (zh) * | 2014-12-03 | 2021-04-13 | 株式会社尼康 | 电子设备 |
Families Citing this family (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6234058B2 (ja) * | 2013-05-09 | 2017-11-22 | キヤノン株式会社 | 焦点調節装置およびその制御方法 |
EP3076662A4 (en) | 2013-11-26 | 2017-08-09 | Nikon Corporation | Electronic device, imaging device, and imaging element |
JP6539015B2 (ja) | 2013-12-05 | 2019-07-03 | キヤノン株式会社 | 撮像装置及びその制御方法 |
JP6198600B2 (ja) * | 2013-12-16 | 2017-09-20 | キヤノン株式会社 | 画像処理装置、撮像装置、その制御方法及びプログラム |
JP6146293B2 (ja) * | 2013-12-25 | 2017-06-14 | ソニー株式会社 | 制御装置、制御方法および制御システム |
JP2015195235A (ja) * | 2014-03-31 | 2015-11-05 | ソニー株式会社 | 固体撮像素子、電子機器、および撮像方法 |
US10485486B2 (en) * | 2014-11-18 | 2019-11-26 | Baylor College Of Medicine | Clinical metric for predicting onset of cardiorespiratory deterioration in patients |
CN105446056B (zh) * | 2014-12-25 | 2018-08-10 | 北京展讯高科通信技术有限公司 | 自动对焦装置及方法 |
KR102328098B1 (ko) * | 2015-02-13 | 2021-11-17 | 삼성전자주식회사 | 전자장치의 초점 검출 장치 및 방법 |
KR20160102814A (ko) * | 2015-02-23 | 2016-08-31 | 삼성전자주식회사 | 이미지 센서 이를 포함하는 이미지 처리 시스템, 및 상기 이미지 센서를 포함하는 모바일 컴퓨팅 장치 |
JP6608185B2 (ja) * | 2015-06-18 | 2019-11-20 | キヤノン株式会社 | 積層型イメージセンサおよび撮像装置 |
KR20180117597A (ko) * | 2016-03-03 | 2018-10-29 | 소니 주식회사 | 화상 처리 장치, 화상 처리 방법, 컴퓨터 프로그램 및 전자 기기 |
CN105827994A (zh) * | 2016-04-12 | 2016-08-03 | 乐视控股(北京)有限公司 | 一种解决对焦区曝光过度的***及方法 |
US10491832B2 (en) * | 2017-08-16 | 2019-11-26 | Qualcomm Incorporated | Image capture device with stabilized exposure or white balance |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2006049361A (ja) | 2004-07-30 | 2006-02-16 | Sony Corp | 半導体モジュール及びmos型固体撮像装置 |
JP2006303755A (ja) * | 2005-04-19 | 2006-11-02 | Konica Minolta Photo Imaging Inc | 撮像装置及び撮像方法 |
JP2009049858A (ja) * | 2007-08-22 | 2009-03-05 | Canon Inc | 撮像装置 |
JP2009094881A (ja) | 2007-10-10 | 2009-04-30 | Nikon Corp | 撮像装置および撮像方法 |
US20100002940A1 (en) | 2008-07-03 | 2010-01-07 | Sony Corporation | Image data processing apparatus and image data processing method |
JP2012060370A (ja) * | 2010-09-08 | 2012-03-22 | Pentax Ricoh Imaging Co Ltd | 撮像システムおよび画素信号読出し方法 |
JP2012139026A (ja) | 2010-12-27 | 2012-07-19 | Alpine Electronics Inc | 消費電力制御システム |
Family Cites Families (41)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH03243071A (ja) * | 1990-02-21 | 1991-10-30 | Sony Corp | オートフォーカス回路 |
JPH06303379A (ja) | 1993-04-14 | 1994-10-28 | Fuji Xerox Co Ltd | 画像読取素子 |
JP3270252B2 (ja) * | 1994-09-12 | 2002-04-02 | オリンパス光学工業株式会社 | カメラ |
JPH0918773A (ja) * | 1995-06-27 | 1997-01-17 | Canon Inc | 撮像装置 |
JP4100736B2 (ja) | 1996-09-27 | 2008-06-11 | シャープ株式会社 | 非水系二次電池用正極活物質ニッケル酸リチウムの製造方法 |
JP2002116369A (ja) * | 2000-10-05 | 2002-04-19 | Ricoh Co Ltd | 自動焦点制御装置 |
JP2003289469A (ja) * | 2002-03-28 | 2003-10-10 | Minolta Co Ltd | 撮像装置 |
JP2003333506A (ja) * | 2002-05-10 | 2003-11-21 | Canon Inc | 画像処理装置、画像処理システム、画像処理方法、プログラム、及び記憶媒体 |
US6747808B2 (en) * | 2002-10-04 | 2004-06-08 | Hewlett-Packard Development Company, L.P. | Electronic imaging device focusing |
JP4665422B2 (ja) | 2004-04-02 | 2011-04-06 | ソニー株式会社 | 撮像装置 |
JP4612814B2 (ja) * | 2004-07-26 | 2011-01-12 | キヤノン株式会社 | 自動焦点調節装置及びその制御方法並びに撮像装置 |
JP2006109103A (ja) | 2004-10-05 | 2006-04-20 | Victor Co Of Japan Ltd | 撮像装置のフォーカス処理回路 |
TW201101476A (en) * | 2005-06-02 | 2011-01-01 | Sony Corp | Semiconductor image sensor module and method of manufacturing the same |
JP4218670B2 (ja) * | 2005-09-27 | 2009-02-04 | オムロン株式会社 | 前方撮影装置 |
JP4816336B2 (ja) * | 2006-02-07 | 2011-11-16 | 日本ビクター株式会社 | 撮像方法及び撮像装置 |
JP2007228019A (ja) * | 2006-02-21 | 2007-09-06 | Olympus Corp | 撮像装置 |
JP2007248782A (ja) * | 2006-03-15 | 2007-09-27 | Olympus Imaging Corp | 焦点調節装置およびカメラ |
US8049801B2 (en) * | 2006-09-14 | 2011-11-01 | Nikon Corporation | Image sensor and imaging apparatus |
US8031258B2 (en) * | 2006-10-04 | 2011-10-04 | Omnivision Technologies, Inc. | Providing multiple video signals from single sensor |
JP5009091B2 (ja) * | 2007-08-23 | 2012-08-22 | 富士フイルム株式会社 | 焦点調節装置及び焦点調節方法 |
JP5056297B2 (ja) * | 2007-09-14 | 2012-10-24 | カシオ計算機株式会社 | 撮像装置、撮像装置制御プログラム及び撮像装置制御方法 |
JP5194688B2 (ja) | 2007-10-01 | 2013-05-08 | 株式会社ニコン | 固体撮像装置 |
JP2009141631A (ja) * | 2007-12-05 | 2009-06-25 | Canon Inc | 光電変換装置及び撮像装置 |
US8441535B2 (en) * | 2008-03-05 | 2013-05-14 | Omnivision Technologies, Inc. | System and method for independent image sensor parameter control in regions of interest |
JP4530067B2 (ja) | 2008-03-27 | 2010-08-25 | ソニー株式会社 | 撮像装置、撮像方法及びプログラム |
US7667169B2 (en) | 2008-05-22 | 2010-02-23 | Omnivision Technologies, Inc. | Image sensor with simultaneous auto-focus and image preview |
JP2010020015A (ja) * | 2008-07-09 | 2010-01-28 | Canon Inc | 撮像装置 |
JP5266916B2 (ja) * | 2008-07-09 | 2013-08-21 | ソニー株式会社 | 撮像素子、カメラ、撮像素子の制御方法、並びにプログラム |
US8179466B2 (en) * | 2009-03-11 | 2012-05-15 | Eastman Kodak Company | Capture of video with motion-speed determination and variable capture rate |
JP5238673B2 (ja) * | 2009-11-09 | 2013-07-17 | 株式会社東芝 | 固体撮像装置 |
KR101624654B1 (ko) * | 2010-03-15 | 2016-05-26 | 삼성전자주식회사 | 초점 조절 장치 및 초점 조절 방법 |
JP5633323B2 (ja) | 2010-11-11 | 2014-12-03 | ソニー株式会社 | 固体撮像装置及び電子機器 |
JP2012129425A (ja) * | 2010-12-16 | 2012-07-05 | Canon Inc | マトリクス基板、検出装置、検出システム、及び、検出装置の駆動方法 |
JP2012175234A (ja) | 2011-02-18 | 2012-09-10 | Sony Corp | 撮像装置、撮像素子、および撮像制御方法、並びにプログラム |
US20130002936A1 (en) * | 2011-06-30 | 2013-01-03 | Nikon Corporation | Image pickup apparatus, image processing apparatus, and storage medium storing image processing program |
JP2013145779A (ja) * | 2012-01-13 | 2013-07-25 | Sony Corp | 固体撮像装置及び電子機器 |
JP2013172210A (ja) * | 2012-02-17 | 2013-09-02 | Canon Inc | 撮像装置 |
JP6023437B2 (ja) * | 2012-02-29 | 2016-11-09 | キヤノン株式会社 | 固体撮像装置及びカメラ |
EP2846538B1 (en) | 2012-05-02 | 2021-06-23 | Nikon Corporation | Imaging device |
WO2014133076A1 (ja) | 2013-02-27 | 2014-09-04 | 株式会社ニコン | 撮像素子および電子機器 |
US9661243B2 (en) * | 2013-03-15 | 2017-05-23 | Forza Silicon Corporation | CMOS image sensor with column-wise selective charge-domain binning |
-
2014
- 2014-02-27 WO PCT/JP2014/054865 patent/WO2014133076A1/ja active Application Filing
- 2014-02-27 CN CN201811159203.XA patent/CN109348148B/zh active Active
- 2014-02-27 JP JP2015503018A patent/JP6090425B2/ja active Active
- 2014-02-27 US US14/771,123 patent/US10924697B2/en active Active
- 2014-02-27 EP EP14756999.0A patent/EP2963917B1/en active Active
- 2014-02-27 CN CN201480010575.XA patent/CN105103537B/zh active Active
-
2017
- 2017-02-09 JP JP2017022351A patent/JP2017118579A/ja active Pending
-
2019
- 2019-10-09 JP JP2019185900A patent/JP2020036330A/ja active Pending
-
2021
- 2021-01-12 US US17/146,870 patent/US11595604B2/en active Active
- 2021-07-20 JP JP2021119878A patent/JP7264189B2/ja active Active
-
2023
- 2023-02-02 US US18/104,936 patent/US20230199343A1/en active Pending
- 2023-04-12 JP JP2023064809A patent/JP2023089131A/ja active Pending
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2006049361A (ja) | 2004-07-30 | 2006-02-16 | Sony Corp | 半導体モジュール及びmos型固体撮像装置 |
JP2006303755A (ja) * | 2005-04-19 | 2006-11-02 | Konica Minolta Photo Imaging Inc | 撮像装置及び撮像方法 |
JP2009049858A (ja) * | 2007-08-22 | 2009-03-05 | Canon Inc | 撮像装置 |
JP2009094881A (ja) | 2007-10-10 | 2009-04-30 | Nikon Corp | 撮像装置および撮像方法 |
US20100002940A1 (en) | 2008-07-03 | 2010-01-07 | Sony Corporation | Image data processing apparatus and image data processing method |
JP2010016621A (ja) | 2008-07-03 | 2010-01-21 | Sony Corp | 画像データ処理装置と画像データ処理方法およびプログラムと記録媒体 |
JP2012060370A (ja) * | 2010-09-08 | 2012-03-22 | Pentax Ricoh Imaging Co Ltd | 撮像システムおよび画素信号読出し方法 |
JP2012139026A (ja) | 2010-12-27 | 2012-07-19 | Alpine Electronics Inc | 消費電力制御システム |
Cited By (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2020036330A (ja) * | 2013-02-27 | 2020-03-05 | 株式会社ニコン | 撮像装置および撮像素子 |
US11595604B2 (en) | 2013-02-27 | 2023-02-28 | Nikon Corporation | Image sensor and imaging device including a plurality of semiconductor substrates |
US10924697B2 (en) | 2013-02-27 | 2021-02-16 | Nikon Corporation | Image sensor and electronic device having imaging regions for focus detection extending in first and second different directions |
EP3203717A4 (en) * | 2014-09-30 | 2018-04-25 | Nikon Corporation | Electronic apparatus, reproduction device, reproduction method, recording medium, and recording method |
EP3203722A4 (en) * | 2014-09-30 | 2018-06-06 | Nikon Corporation | Electronic apparatus and image processing method |
JPWO2016052640A1 (ja) * | 2014-09-30 | 2017-08-31 | 株式会社ニコン | 電子機器、撮像装置、および、撮像システム |
US11689798B2 (en) | 2014-09-30 | 2023-06-27 | Nikon Corporation | Electronic device, image-capturing device, and image-capturing system |
EP3203723A4 (en) * | 2014-09-30 | 2018-02-28 | Nikon Corporation | Electronic device |
EP3203720A4 (en) * | 2014-09-30 | 2018-03-21 | Nikon Corporation | Electronic apparatus, reproduction device, reproduction method, recording medium, and recording method |
EP3203718A4 (en) * | 2014-09-30 | 2018-04-11 | Nikon Corporation | Electronic apparatus, reproduction device, reproduction method, recording medium, and recording method |
WO2016052640A1 (ja) * | 2014-09-30 | 2016-04-07 | 株式会社ニコン | 電子機器、撮像装置、および、撮像システム |
EP3203721A4 (en) * | 2014-09-30 | 2018-05-16 | Nikon Corporation | Electronic apparatus, image pickup device, and image pickup system |
CN107005636A (zh) * | 2014-09-30 | 2017-08-01 | 株式会社尼康 | 电子设备 |
CN107005633A (zh) * | 2014-09-30 | 2017-08-01 | 株式会社尼康 | 电子设备、再现装置、再现方法、记录介质及记录方法 |
JP2020065280A (ja) * | 2014-09-30 | 2020-04-23 | 株式会社ニコン | 電子機器 |
CN107005638A (zh) * | 2014-09-30 | 2017-08-01 | 株式会社尼康 | 电子设备、拍摄装置和拍摄*** |
CN107005635A (zh) * | 2014-09-30 | 2017-08-01 | 株式会社尼康 | 电子设备、再现装置、再现方法、记录介质及记录方法 |
JP2022031703A (ja) * | 2014-09-30 | 2022-02-22 | 株式会社ニコン | 電子機器 |
US11006067B2 (en) | 2014-09-30 | 2021-05-11 | Nikon Corporation | Electronic apparatus, reproduction device, reproduction method, recording medium, and recording method |
CN113890993A (zh) * | 2014-09-30 | 2022-01-04 | 株式会社尼康 | 拍摄元件和拍摄装置 |
CN112653821A (zh) * | 2014-12-03 | 2021-04-13 | 株式会社尼康 | 电子设备 |
JP2017225084A (ja) * | 2016-06-17 | 2017-12-21 | 株式会社ニコン | 撮像素子、電子機器、撮像方法、およびプログラム |
JP2020101969A (ja) * | 2018-12-21 | 2020-07-02 | 株式会社モルフォ | 画像処理装置、画像処理方法およびプログラム |
Also Published As
Publication number | Publication date |
---|---|
US20230199343A1 (en) | 2023-06-22 |
CN105103537A (zh) | 2015-11-25 |
EP2963917A4 (en) | 2016-10-19 |
JP6090425B2 (ja) | 2017-03-08 |
EP2963917B1 (en) | 2021-04-14 |
EP2963917A1 (en) | 2016-01-06 |
JPWO2014133076A1 (ja) | 2017-02-02 |
US10924697B2 (en) | 2021-02-16 |
CN109348148B (zh) | 2022-03-15 |
JP7264189B2 (ja) | 2023-04-25 |
CN109348148A (zh) | 2019-02-15 |
JP2023089131A (ja) | 2023-06-27 |
JP2021168512A (ja) | 2021-10-21 |
JP2020036330A (ja) | 2020-03-05 |
CN105103537B (zh) | 2018-11-06 |
US20160014359A1 (en) | 2016-01-14 |
US20210136302A1 (en) | 2021-05-06 |
JP2017118579A (ja) | 2017-06-29 |
US11595604B2 (en) | 2023-02-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP7264189B2 (ja) | 撮像素子及び撮像装置 | |
JP6780745B2 (ja) | 電子機器 | |
WO2014192152A1 (ja) | 電子機器及び制御プログラム | |
JP2014178603A (ja) | 撮像装置 | |
JP6413233B2 (ja) | 撮像装置および撮像素子 | |
US10757341B2 (en) | Electronic device, imaging device, and imaging element | |
JP2014179892A (ja) | 撮像装置 | |
JP2014179778A (ja) | 信号処理装置、撮像素子、撮像装置および電子機器 | |
JP6488545B2 (ja) | 電子機器 | |
JP6781140B2 (ja) | 撮像装置 | |
JP6349659B2 (ja) | 電子機器、電子機器の制御方法、及び制御プログラム | |
JP6852712B2 (ja) | 撮像素子および撮像装置 | |
JP6916418B2 (ja) | 撮像装置 | |
JP2021193811A (ja) | 撮像素子、及び撮像装置 | |
JP2018174543A (ja) | 電子機器 | |
JP2019092220A (ja) | 電子機器 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201480010575.X Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14756999 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2015503018 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2014756999 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14771123 Country of ref document: US |