CN106454090A - Automatic focusing method and system based on depth camera - Google Patents

Automatic focusing method and system based on depth camera Download PDF

Info

Publication number
CN106454090A
CN106454090A CN201610886040.XA CN201610886040A CN106454090A CN 106454090 A CN106454090 A CN 106454090A CN 201610886040 A CN201610886040 A CN 201610886040A CN 106454090 A CN106454090 A CN 106454090A
Authority
CN
China
Prior art keywords
depth
focusing
camera
image
focusing area
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610886040.XA
Other languages
Chinese (zh)
Other versions
CN106454090B (en
Inventor
黄源浩
刘龙
肖振中
许星
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Orbbec Co Ltd
Original Assignee
Shenzhen Orbbec Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Orbbec Co Ltd filed Critical Shenzhen Orbbec Co Ltd
Priority to CN201610886040.XA priority Critical patent/CN106454090B/en
Publication of CN106454090A publication Critical patent/CN106454090A/en
Application granted granted Critical
Publication of CN106454090B publication Critical patent/CN106454090B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/67Focus control based on electronic image sensor signals
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B7/00Mountings, adjusting means, or light-tight connections, for optical elements
    • G02B7/28Systems for automatic generation of focusing signals
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03BAPPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
    • G03B13/00Viewfinders; Focusing aids for cameras; Means for focusing for cameras; Autofocus systems for cameras
    • G03B13/32Means for focusing
    • G03B13/34Power focusing
    • G03B13/36Autofocus systems

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Optics & Photonics (AREA)
  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Studio Devices (AREA)

Abstract

The invention relates to the field of automatic focusing, in particular to a method for realizing automatic focusing by acquisition of depth information through a depth camera and a camera system. The method comprises the following steps: S1, acquiring a non-depth image of a target area by using a focusing camera; S2, acquiring a focusing area needing to be focused in the non-depth image; S3, intercepting a depth extraction area only including the focusing area at a corresponding position on an acquired image by the depth camera according to the focusing area, only acquiring a depth image of the depth extraction area by the depth camera, and calculating depth information of the focusing area according to the depth extraction area; and S4, performing automatic focusing by the focusing camera according to the depth information. Through adoption of the method and the system, rapid and high-accuracy automatic focusing can be realized at any position in a target space.

Description

Atomatic focusing method based on depth camera and system
Technical field
The present invention relates to auto-focusing field, more particularly to one kind obtain depth information to realize certainly using depth camera The method of dynamic focusing and camera system.
Background technology
Autofocus Technology is to realize the key that camera shoots high quality graphic, and Autofocus Technology is broadly divided into passive type With active auto-focusing.Passive type auto-focusing is to be carried out using being focused the definition in image formed by image planes for the object Focusing, usually, needs to be iterated through climbing method or gradient decline scheduling algorithm, that is, the adjusting focal length passing through each is straight It is the most clearly to this method advantage is not need using other devices, and shortcoming is that focusing speed is slower to image.Main Dynamic formula auto-focusing is to realize focusing according to the depth value being focused object distance camera lens, for the side of Measuring Object depth value Method has laser ranging method, supersonic sounding method etc., and the advantage of this method is quickly carry out auto-focusing, and weak point exists To fathom in the device needing extra.
With the increase of photographic head quantity in recent years, the appearance of such as dual camera, active Autofocus Technology goes out again Show and calculated depth using binocular vision method and then realize focusing.It is true that due to binocular vision calculate depth algorithm relatively For complexity, and precision it cannot be guaranteed that, thus this Autofocus Technology nor reach good effect.
The current active focusing technology based on depth camera, the acquisition for depth information is usually to obtain target Just complete the acquisition of the depth information of all depth images before region, thus there is a problem in that, due to needing in advance Depth information calculating carried out to entire image, and to store all of depth information, lead to current Autofocus Technology to calculate Amount is very big, particularly, under the dynamic mode similar to photography etc., since it is desired that real-time ensuring image is clear, so The process of focusing needs to run always, and minimum frame per second also want 20-30FPS at present, within the vision under each frame Each pixel of image completes the acquisition of depth information it is clear that this amount of calculation is very big, if for higher frame per second, this Amount of calculation will be bigger, so focusing speed will become very slow, the image definition directly resulting in shooting reduces.
In addition, the above-mentioned Autofocus Technology based on depth camera, due to computationally intensive, for multi-spot more Can't bear the heavy load, so not enabling the function of multi-spot.
The disclosure of background above technology contents is only used for assisting the inventive concept understanding the present invention and technical scheme, and it is not Necessarily belong to the prior art of present patent application, show the applying date in present patent application for the above there is no tangible proof In the case of disclosed, above-mentioned background technology should not be taken to evaluate novelty and the creativeness of the application.
Content of the invention
Present invention aim at proposing a kind of Atomatic focusing method based on depth camera and system, above-mentioned existing to solve Computationally intensive, auto-focusing that cannot be smooth very well the technical problem that what technology existed obtain depth information is improper to be led to.
For this reason, the present invention proposes a kind of Atomatic focusing method based on depth camera, comprise the following steps:
S1:Obtain the non-depth image of target area using zoom camera;
S2:Obtain the focusing area needing to focus in this non-depth image;
S3:Described depth camera, according to this focusing area, on the correspondence position of collection image, intercepts and only includes this focusing The depth extraction region in region, only obtains the depth image in this depth extraction region using depth camera, according to this depth extraction Region calculates the depth information of this focusing area;
S4:Auto-focusing is carried out by described zoom camera according to described depth information.
Preferably, the method that the present invention provides can also have following feature or step:
The size in described depth extraction region is more than the size of described focusing area.
In addition, in order to solve the problems, such as multi-spot, in step S2, this non-depth image includes one or more Need focusing focusing area.
Meanwhile, for the problem of multi-spot, in step S3, described depth camera is according to this focusing area more than one Domain, on the corresponding more than one position of collection image, the depth only including this focusing area intercepting corresponding number carries Take region, obtain the depth image in each described depth extraction region using depth camera respectively, according to each described depth extraction area Domain calculates multiple depth informations of each described focusing area, in step S4, passes through described zoom according to the plurality of depth information Camera carries out auto-focusing.
For single-point focusing, in step S3, described depth camera according to this focusing area, in the correspondence position of collection image On, intercept a depth extraction region only including this focusing area, only obtain this depth extraction region using depth camera Depth image, calculates the depth information of this focusing area according to this depth extraction region.
In above-mentioned steps S3, comprise the following steps:
S31:Obtain the depth image in described depth extraction region using described depth camera;
S32:Described depth camera and described zoom camera are carried out registering, obtain each picture of focusing area in non-depth image The depth value of element;
S33:Calculate the depth information in this region according to the depth value of each pixel of described focusing area.
After obtaining the depth value of each pixel in described focusing area, if focusing area is single pixel point, by this pixel The depth value of point carries out auto-focusing directly as depth information;
If comprising multiple pixels in described focusing area, need for the depth value of multiple pixels to be fused into single depth Value, obtains the single depth information in this region.
By depth value size distribution situation, the depth value choosing the pixel of intermediate distribution is average to do, thus it is right to obtain this The single depth information in burnt region.
For dynamic object and static object, in step S2, described focusing area includes dynamic focusing region and static state is right Burnt region, in step S3, the state according to described focusing area intercepts depth extraction region to the corresponding of collection image.
In addition the invention allows for a kind of autofocus system based on depth camera, using described in any of the above-described Atomatic focusing method executes focusing, including depth camera unit, zoom camera unit, processor unit and controller unit, Depth camera unit and zoom camera unit obtain the depth image of focusing area and non-depth image respectively;Processor list Unit is for obtaining the depth information of the focusing area in non-depth image and this focusing area;Controller unit is used for according to depth Degree information carries out auto-focusing to zoom camera.
The beneficial effect that the present invention is compared with the prior art includes:The Atomatic focusing method of the present invention is to obtain target area The non-depth image in domain, and acquisition is just obtained deeply to using depth camera after needing the focusing area focused in non-depth image Spend image, and be used for extracting the depth extraction region of depth image and only include the region that focusing area is located, intercept On collection image, other regions in addition to focusing area, to obtain depth information with this, so believe compared to existing depth Breath acquisition methods, on the one hand, pass through to obtain the first and second target images through a set of complexity in contrast to using binocular vision method Algorithm to obtain depth value for, this method can be realized quick and high-precision automatically right to any position in object space Jiao, better than the method for active auto-focusing at present, is better than the method based on the passive type auto-focusing such as laser ranging, separately simultaneously On the one hand, the amount of calculation of auto-focusing substantially reduces, and need not all calculate whole depth informations in advance, the impact of frame per second is relatively Little, even if based on frame per second higher in the case of, also can rapid focus, particularly, the present invention is more applicable for the fast of dynamic object Speed focusing, implement focusing effect greatly improve so that the image shooting is relatively sharp, can smoothness auto-focusing.
Brief description
Fig. 1 is the system block diagram of the specific embodiment of the invention one;
Fig. 2 is autofocus system schematic diagram on mobile terminal in the specific embodiment of the invention one;
Fig. 3 is the flow chart of the specific embodiment of the invention two;
Fig. 4 is the flow chart of the specific embodiment of the invention three;
Fig. 5 is the light path schematic diagram of the specific embodiment of the invention two and three.
Fig. 6 is the non-depth image schematic diagram that in the specific embodiment of the invention two, zoom camera unit obtains.
Fig. 7 is that the depth image of the collection module acquisition of depth camera unit in the specific embodiment of the invention two is illustrated Figure.
Fig. 8 is the schematic diagram of registration and auto-focusing in the specific embodiment of the invention two and three.
Fig. 9 is the specific embodiment of the invention two to defocused non-depth image.
Figure 10 is the non-depth image schematic diagram of the zoom camera unit acquisition of the specific embodiment of the invention three.
Figure 11 is that the depth image of the collection module acquisition of the depth camera unit of the specific embodiment of the invention three is illustrated Figure.
Figure 12 is the specific embodiment of the invention three to defocused non-depth image.
Specific embodiment
With reference to specific embodiment and compare accompanying drawing the present invention is described in further detail.It is emphasized that The description below is merely exemplary, rather than in order to limit the scope of the present invention and its application.
With reference to the following drawings, the embodiment of non-limiting and nonexcludability will be described, wherein identical reference represents Identical part, unless stated otherwise.
Embodiment one:
The present embodiment proposes a kind of autofocus system based on depth camera, as shown in figure 1, including depth camera list Unit 1, zoom camera unit 3, processor unit 2 and controller unit 4, depth camera unit 1 and 3 points of zoom camera unit Huo Qu not the depth image of focusing area and non-depth image;Processor unit 2 is used for obtaining the focusing in non-depth image Region and the depth information of this focusing area;Controller unit 4 is automatically right for being carried out to zoom camera according to depth information Burnt.
In the present embodiment, depth camera unit 1 can be the depth camera based on structure light principle or based on TOF principle.
Depth camera based on structure light principle is typically made up of laser projections module and collection module 12, by projection Module, to the encoded structured light patterns of object space projection, gathers the knot after module 12 collects by space object depth modulation Structure light image, recycles trigonometric calculations to go out the actual grade value of each pixel in image.
Camera based on TOF principle is typically made up of transmitter module and receiver module, and wherein transmitter module connects to target 5 Supervention penetrates light pulse, and receives the record light pulse flight time in space after corresponding light pulse by receiver module, Pass through the depth information that conversion obtains target object afterwards.
In the present embodiment, as shown in Fig. 2 being the mobile terminal A being provided with the depth camera based on structure light, throw including laser Penetrate module and collection module 12 and zoom camera, three is in that yi word pattern arranges.
By imageing sensor, such as CCD or CMOS, and zoom lens forms zoom camera unit 3.In addition, in image Can be realized photosensitive to light sources with different wavelengths, thus clapping between sensor and zoom lens by the different optical filters of increase Take out the image of different colours.Such as, when not increasing optical filter or transparent filter, the image of shooting is black white image;When During for Baeyer optical filter, shooting is color RGB image;When for infrared or ultraviolet filter, shooting is infrared or ultraviolet Image;Can also be the combination of multiple optical filters, such as RGB-IR optical filter, comprise in the image now shooting red, blue, green and Infrared information, can obtain RGB image and infrared image after interpolation.
Processor unit 2 is processed by the data that the combination of chip and application realizes camera is collected. It should be noted that turning further for the data in depth camera unit 1 directly obtaining collection module 12 or receiver module The process of chemical conversion depth data is also required to the process of processor, this processor can be in addition special chip it is also possible to by Present processor unit 2 completes, and this technology is known technology, will no longer illustrate.
For this automatic focusing camera system, processor unit 2 is used for obtaining needs focusing in non-depth image Region and the single depth information in this region.
Embodiment two:
In the present embodiment, as shown in figure 3, the autofocus system based on depth camera being proposed according to embodiment one, carry Go out a kind of Atomatic focusing method based on depth camera, the present embodiment is mainly used in single-point focusing, can refer to light path Fig. 5.
First, S1, using zoom camera obtain target area non-depth image 8.
As shown in fig. 6, zoom camera obtains non-depth image 8, the target 5 in non-depth image 8 at this moment can be State or static state, mainly depending on the mode of operation of camera.
Secondly, S2, obtain the focusing area 7 needing to focus in this non-depth image 8;
Usually, in non-depth image 8, moving object is often defaulted as needing the object of focusing, such as human or animal;? Often by zone line focusing area 7 by default;Therefore, it is possible to arrange out the pixel region of acquiescence focusing by program.In addition, Such as can also be touched by finger and obtain focusing area 7 by way of image-region by way of artificial actively selection.Selected In region typically there is certain size and shape, equally can with program acquiescence or picking, such as square, circular etc., also may be used To be a specifically pixel, it is not limited here.
In the present embodiment, under the pattern of single-point focusing, described non-depth image 8 includes the focusing area of a needs focusing Domain 7, and described focusing area 7 includes dynamic focusing region and static focusing area, the state according to target obtains focusing area 7.
Secondly, as shown in fig. 7, S3, described depth camera are according to this focusing area, on the correspondence position of collection image 9, Intercept a depth extraction region 6 only including this focusing area, only obtain the depth in this depth extraction region 6 using depth camera Degree image, calculates the depth information of this focusing area according to this depth extraction region 6.
Due to only needing the depth information of focusing area, thus depth camera need not be allowed to obtain the depth map in whole space Picture.But then, because depth camera and zoom camera are on different positions, specifically, can by depth camera with Zoom camera be arranged in parallel it is also possible to be oppositely arranged at an angle, and this results in depth camera and zoom camera exists Parallax, that is, the residing location of pixels on depth image with non-depth image 8 of the same space point differ.But in advance due to not knowing Road each one-to-one relationship between pixel, thus be accomplished by obtaining the depth bigger than focusing area using depth camera and carry Take the depth image in region 6 it is ensured that all pixels have corresponding depth value in depth extraction region 6 in focusing area. Next, for the depth value obtaining each pixel in focusing area in non-depth image 8, needing first to depth camera and zoom phase Machine carries out registration.
As shown in figure 8, registering, that is, utilize depth camera and zoom camera the outer ginseng of relative position relation composition and The camera internal reference that each parameter forms, finds depth camera and each picture of zoom camera using pinhole camera model and trigonometry The one-to-one relationship of element.Specifically registering principle is prior art, and here is not detailed.After registration, need in non-depth image 8 Want the depth value equal to respective pixel in depth image for the depth value of each pixel on focusing area.
Finally, S4, auto-focusing is carried out by described zoom camera according to described depth information.As shown in figure 8,10 is right Burnt area image, 13 in Fig. 9 are the final non-depth image of single-point focusing.
After obtaining the depth value of each pixel in focusing area, if focusing area is single pixel point, the depth of this pixel Value just can be used directly to carry out auto-focusing.If comprising multiple pixels in focusing area, need multiple pixels Depth value is fused into single depth value.Preferably, the meansigma methodss taking the depth value of each pixel in this region are as this region Single depth information;Further, in order to avoid respective pixel depth value is too big or too little thus in influence area focus object Accurate depth, by depth value size distribution situation, the depth value choosing the pixel of intermediate distribution is average to do, thus being somebody's turn to do The single depth information in region.Can also be obtained by additive method, be not limited thereto.
Need exist for illustrating, if wanting to realize the focusing in all pixels region in non-depth image it is necessary to obtain The depth value of all pixels, requires the FOV of depth camera should be not less than the FOV of focusing camera for this.In one embodiment, depth The camera lens of magazine receiving module is wide-angle lens, and the telephoto lens of the camera lens of focusing camera.So, described depth phase The angle of visual field of machine is not less than the angle of visual field of zoom camera.
Controller unit is used for carrying out auto-focusing according to depth information to zoom camera.Need focusing area obtaining Single depth information after, adjustment zoom camera camera lens focal length with meet to this depth focus on.Adjustment can be by advance The program setting, to complete, specifically, has certain relation, by this relation with the shape of program between focal length size and depth size Formula is saved in the storage of camera system, after getting single depth information, calculates the amount of adjustment according to program, then Auto-focusing is realized by controller unit.Memory cell described here almost all exists in all of electronic equipment, all For data storage, thus it is present in the system by acquiescence, be not described in detail in here.
The Atomatic focusing method of the present embodiment is in the non-depth image obtaining target area, and obtains in non-depth image Just obtain depth image to using depth camera after taking the focusing area needing to focus, and be used for extracting the depth of depth image Degree extracts the region that region 6 only includes focusing area place, has intercepted other areas in addition to focusing area on collection image 9 Domain, to obtain depth information with this, so compared to existing depth information acquisition method, on the one hand, in contrast to adopting binocular , by obtaining the first and second target images for a set of complicated algorithm is to obtain depth value, this method is permissible for visual method Any position in object space is realized with quick and high-precision auto-focusing, better than the method for active auto-focusing at present, It is better than the method based on the passive type auto-focusing such as laser ranging, on the other hand, the amount of calculation of auto-focusing substantially reduces simultaneously, Whole depth informations need not all be calculated in advance, the impact of frame per second is less, even if based on frame per second higher in the case of, also can Rapid focus, particularly, the present invention is more applicable for the rapid focus of dynamic object, and the effect implementing focusing greatly improves, and makes The image that must shoot is relatively sharp, auto-focusing that can be smooth.
Embodiment three:
In the present embodiment, as shown in figure 4, the autofocus system based on depth camera being proposed according to embodiment one, carry Go out a kind of Atomatic focusing method based on depth camera, with the difference of embodiment two, the present embodiment is that the present embodiment is main For multi-spot.
As shown in Figure 10, first, S1, using zoom camera obtain target area non-depth image 14.
Zoom camera obtains non-depth image, the target in non-depth image at this moment can be dynamic or Static, mainly depending on the mode of operation of camera.
Secondly, S2, obtain the focusing area needing to focus in this non-depth image 14;
Usually, in non-depth image, moving object is often defaulted as needing the object of focusing, such as human or animal;Also normal By zone line focusing area by default;Therefore, it is possible to arrange out the pixel region of acquiescence focusing by program.In addition, Such as can be touched by finger and obtain focusing area by way of image-region by way of artificial actively selection.Selected Region typically there is certain size and shape, equally can with program acquiescence or picking, such as square, circular etc. it is also possible to It is a specifically pixel, be not limited here.
In the present embodiment, under the pattern of multi-spot, described non-depth image includes the right of more than one needs focusing Burnt region, described focusing area also can be divided into dynamic focusing region and static focusing area, and the state according to target obtains focusing Region.
Secondly, as shown in figure 11, S3, described depth camera, according to this focusing area more than one, are gathering image 9 On corresponding more than one position, intercept the depth extraction region only including this focusing area of corresponding number, using depth Camera obtains the depth image 15 in each described depth extraction region respectively, and it is each described right to be calculated according to each described depth extraction region Multiple depth informations in burnt region.
Due to only needing the depth information of focusing area, thus depth camera need not be allowed to obtain the depth map in whole space Picture.But then, because depth camera and zoom camera are on different positions, specifically, can by depth camera with Zoom camera be arranged in parallel it is also possible to be oppositely arranged at an angle, and this results in depth camera and zoom camera exists Parallax, that is, the residing location of pixels on depth image with non-depth image of the same space point differ.But in advance due to not knowing Road each one-to-one relationship between pixel, thus be accomplished by obtaining the depth bigger than focusing area using depth camera and carry Take the depth image in region it is ensured that all pixels have corresponding depth value in depth extraction region in focusing area.Connect Get off, in order to obtain the depth value of each pixel in focusing area in non-depth image, need first to depth camera and zoom camera Carry out registration.
Registration, is each joined using the outer ginseng and camera of the relative position relation composition of depth camera and zoom camera The internal reference that array becomes, find depth camera and each pixel of zoom camera using pinhole camera model and trigonometry one a pair Should be related to.Specifically registering principle is prior art, and here is not detailed.After registration, in non-depth image, need focusing area The depth value of upper each pixel is equal to the depth value of respective pixel in depth image.
Finally, S4, auto-focusing is carried out by described zoom camera according to the plurality of depth information.As shown in figure 12, 16 is the final non-depth image of multi-spot.
After obtaining the depth value of each pixel in focusing area, if focusing area is single pixel point, the depth of this pixel Value just can be used directly to carry out auto-focusing.If comprising multiple pixels in focusing area, need multiple pixels Depth value is fused into single depth value.Preferably, the meansigma methodss taking the depth value of each pixel in this region are as this region Single depth information;Further, in order to avoid respective pixel depth value is too big or too little thus in influence area focus object Accurate depth, by depth value size distribution situation, the depth value choosing the pixel of intermediate distribution is average to do, thus being somebody's turn to do The single depth information in region.Can also be obtained by additive method, be not limited thereto.
Need exist for illustrating, if wanting to realize the focusing in all pixels region in non-depth image it is necessary to obtain The depth value of all pixels, requires the FOV of depth camera should be not less than the FOV (angle of visual field) of focusing camera for this.One enforcement In example, the camera lens of the receiving module in depth camera is wide-angle lens, and the telephoto lens of the camera lens of focusing camera.So, institute The angle of visual field of the depth camera stated is not less than the angle of visual field of zoom camera.
Controller unit is used for carrying out auto-focusing according to depth information to zoom camera.Need focusing area obtaining Single depth information after, adjustment zoom camera camera lens focal length with meet to this depth focus on.Adjustment can be by advance The program setting, to complete, specifically, has certain relation, by this relation with the shape of program between focal length size and depth size Formula is saved in the storage of camera system, after getting single depth information, calculates the amount of adjustment according to program, then Auto-focusing is realized by controller unit.Memory cell described here almost all exists in all of electronic equipment, all For data storage, thus it is present in the system by acquiescence, be not described in detail in here.
In this enforcement, due to using be intercept after the depth information in depth extraction region carry out auto-focusing, so Can realize carrying out the possibility of multi-spot using depth camera, during focusing, due to only obtaining more than one Focusing area more than one depth image, the region beyond focusing area do not carry out the acquisition of depth value, even if so It is that multiple points are carried out with auto-focusing amount of calculation is also smaller, the quick calculating under current frame rate can be met, and then submit to real Existing rapid focus, and the current Atomatic focusing method based on depth camera, general, it is by depth whole for target area Image obtains in advance, and all of depth information is realized storage, can not quickly find focusing for multi-spot accurately The depth information in region, especially under dynamic mode, because focus process is being carried out always, so, realize quickly Multi-spot is even more impossible, and general frame per second also has 20-30FPS in addition, thus big the thinking of amount of calculation of multi-spot and Know.
Using the depth information of the present embodiment acquisition method it is achieved that multiple spot auto-focusing based on depth camera can Can, focusing speed is fast in addition, and amount of calculation is little, has significantly heightened focusing speed and practicality.
It would be recognized by those skilled in the art that above description is made with numerous accommodations is possible, so embodiment is only It is used for describing one or more particular implementation.
Although having been described above and describe the example embodiment being counted as the present invention, it will be apparent to those skilled in the art that It can be variously modified and replace, without departing from the spirit of the present invention.Furthermore it is possible to make many modifications with by spy Condition of pledging love is fitted to the religious doctrine of the present invention, without departing from invention described herein central concept.So, the present invention is unrestricted In specific embodiment disclosed here, but the present invention may also include belonging to all embodiments of the scope of the invention and its be equal to Thing.

Claims (10)

1. a kind of Atomatic focusing method based on depth camera is it is characterised in that comprise the following steps:
S1:Obtain the non-depth image of target area using zoom camera;
S2:Obtain the focusing area needing to focus in this non-depth image;
S3:Described depth camera, according to this focusing area, on the correspondence position of collection image, intercepts and only includes this focusing area Depth extraction region, only obtain the depth image in this depth extraction region using depth camera, according to this depth extraction region Calculate the depth information of this focusing area;
S4:Auto-focusing is carried out by described zoom camera according to described depth information.
2. Atomatic focusing method as claimed in claim 1 it is characterised in that:The size in described depth extraction region is more than described The size of focusing area.
3. Atomatic focusing method as claimed in claim 1 it is characterised in that:In step S2, this non-depth image includes one The individual or more than one focusing area needing focusing.
4. Atomatic focusing method as claimed in claim 3 it is characterised in that:In step S3, described depth camera is according to one This above focusing area, on the corresponding more than one position of collection image, intercept corresponding number only to include this right The depth extraction region in burnt region, obtains the depth image in each described depth extraction region respectively, according to each using depth camera Described depth extraction region calculates multiple depth informations of each described focusing area, in step S4, is believed according to the plurality of depth Breath carries out auto-focusing by described zoom camera.
5. Atomatic focusing method as claimed in claim 3 it is characterised in that:In step S3, described depth camera is right according to this Burnt region, on the correspondence position of collection image, intercepts a depth extraction region only including this focusing area, using depth Camera only obtains the depth image in this depth extraction region, calculates the depth letter of this focusing area according to this depth extraction region Breath.
6. the Atomatic focusing method as described in any one of claim 1-5 it is characterised in that:In step S3, comprise the following steps:
S31:Obtain the depth image in described depth extraction region using described depth camera;
S32:Described depth camera and described zoom camera are carried out registering, obtain each pixel of focusing area in non-depth image Depth value;
S33:Calculate the depth information in this region according to the depth value of each pixel of described focusing area.
7. Atomatic focusing method as claimed in claim 6 it is characterised in that:Obtain the depth of each pixel in described focusing area After value, if focusing area is single pixel point, the depth value of this pixel is carried out auto-focusing directly as depth information;
If comprising multiple pixels in described focusing area, need for the depth value of multiple pixels to be fused into single depth value, Obtain the single depth information in this region.
8. Atomatic focusing method as claimed in claim 6 it is characterised in that:By depth value size distribution situation, choose middle The depth value of the pixel of distribution is average to do, thus obtaining the single depth information of this focusing area.
9. Atomatic focusing method as claimed in claim 1 it is characterised in that:In step S2, described focusing area includes dynamically Focusing area and static focusing area, in step S3, the state according to described focusing area is deep to the corresponding intercepting of collection image Degree extracts region.
10. a kind of autofocus system based on depth camera it is characterised in that:Using described in any one of claim 1-9 Atomatic focusing method executes focusing, including depth camera unit, zoom camera unit, processor unit and controller unit, Depth camera unit and zoom camera unit obtain the depth image of focusing area and non-depth image respectively;Processor list Unit is for obtaining the depth information of the focusing area in non-depth image and this focusing area;Controller unit is used for according to depth Degree information carries out auto-focusing to zoom camera.
CN201610886040.XA 2016-10-09 2016-10-09 Atomatic focusing method and system based on depth camera Active CN106454090B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610886040.XA CN106454090B (en) 2016-10-09 2016-10-09 Atomatic focusing method and system based on depth camera

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610886040.XA CN106454090B (en) 2016-10-09 2016-10-09 Atomatic focusing method and system based on depth camera

Publications (2)

Publication Number Publication Date
CN106454090A true CN106454090A (en) 2017-02-22
CN106454090B CN106454090B (en) 2019-04-09

Family

ID=58173351

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610886040.XA Active CN106454090B (en) 2016-10-09 2016-10-09 Atomatic focusing method and system based on depth camera

Country Status (1)

Country Link
CN (1) CN106454090B (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107959799A (en) * 2017-12-18 2018-04-24 信利光电股份有限公司 A kind of quick focusing method, device, equipment and computer-readable recording medium
CN108055458A (en) * 2017-12-18 2018-05-18 信利光电股份有限公司 A kind of focus method for tracing, device, equipment and computer readable storage medium
CN109357628A (en) * 2018-10-23 2019-02-19 北京的卢深视科技有限公司 The high-precision three-dimensional image-pickup method and device of area-of-interest
WO2019037088A1 (en) * 2017-08-25 2019-02-28 深圳市大疆创新科技有限公司 Exposure control method and device, and unmanned aerial vehicle
CN109831609A (en) * 2019-03-05 2019-05-31 上海炬佑智能科技有限公司 TOF depth camera and its Atomatic focusing method
CN109901134A (en) * 2017-12-08 2019-06-18 浙江舜宇智能光学技术有限公司 Wide-angle TOF mould group and its application
CN110262055A (en) * 2019-05-24 2019-09-20 深圳奥比中光科技有限公司 Projective module group, depth camera and the method for adjusting projective module group focal length
WO2020042000A1 (en) * 2018-08-29 2020-03-05 深圳市大疆创新科技有限公司 Camera device and focusing method
WO2020087485A1 (en) * 2018-11-02 2020-05-07 Oppo广东移动通信有限公司 Method for acquiring depth image, device for acquiring depth image, and electronic device
CN112328150A (en) * 2020-11-18 2021-02-05 贝壳技术有限公司 Automatic screenshot method, device and equipment, and storage medium
CN112749610A (en) * 2020-07-27 2021-05-04 腾讯科技(深圳)有限公司 Depth image, reference structured light image generation method and device and electronic equipment
CN113382165A (en) * 2021-05-31 2021-09-10 上海集成电路制造创新中心有限公司 Camera focusing method and camera focusing system
CN113466339A (en) * 2021-06-18 2021-10-01 之江实验室 Ultrasonic scanning microscope global focusing method and device combined with depth camera
CN115037880A (en) * 2022-07-13 2022-09-09 山西工程职业学院 Quick focusing method for airborne camera
CN115379108A (en) * 2018-11-20 2022-11-22 中山市远尚光电科技有限公司 Telephoto lens based on automatic distance measurement

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105120257A (en) * 2015-08-18 2015-12-02 宁波盈芯信息科技有限公司 Vertical depth sensing device based on structured light coding
CN105744163A (en) * 2016-02-23 2016-07-06 湖南拓视觉信息技术有限公司 Video camera and video recording method performing tracking focusing based on depth information
CN105744138A (en) * 2014-12-09 2016-07-06 联想(北京)有限公司 Quick focusing method and electronic equipment
US20160261844A1 (en) * 2015-03-06 2016-09-08 Massachusetts Institute Of Technology Methods and Apparatus for Enhancing Depth Maps with Polarization Cues

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105744138A (en) * 2014-12-09 2016-07-06 联想(北京)有限公司 Quick focusing method and electronic equipment
US20160261844A1 (en) * 2015-03-06 2016-09-08 Massachusetts Institute Of Technology Methods and Apparatus for Enhancing Depth Maps with Polarization Cues
CN105120257A (en) * 2015-08-18 2015-12-02 宁波盈芯信息科技有限公司 Vertical depth sensing device based on structured light coding
CN105744163A (en) * 2016-02-23 2016-07-06 湖南拓视觉信息技术有限公司 Video camera and video recording method performing tracking focusing based on depth information

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019037088A1 (en) * 2017-08-25 2019-02-28 深圳市大疆创新科技有限公司 Exposure control method and device, and unmanned aerial vehicle
CN109901134A (en) * 2017-12-08 2019-06-18 浙江舜宇智能光学技术有限公司 Wide-angle TOF mould group and its application
CN108055458A (en) * 2017-12-18 2018-05-18 信利光电股份有限公司 A kind of focus method for tracing, device, equipment and computer readable storage medium
CN107959799A (en) * 2017-12-18 2018-04-24 信利光电股份有限公司 A kind of quick focusing method, device, equipment and computer-readable recording medium
CN111345025A (en) * 2018-08-29 2020-06-26 深圳市大疆创新科技有限公司 Camera device and focusing method
WO2020042000A1 (en) * 2018-08-29 2020-03-05 深圳市大疆创新科技有限公司 Camera device and focusing method
CN109357628A (en) * 2018-10-23 2019-02-19 北京的卢深视科技有限公司 The high-precision three-dimensional image-pickup method and device of area-of-interest
WO2020087485A1 (en) * 2018-11-02 2020-05-07 Oppo广东移动通信有限公司 Method for acquiring depth image, device for acquiring depth image, and electronic device
US11494925B2 (en) 2018-11-02 2022-11-08 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Method for depth image acquisition, electronic device, and storage medium
CN115379108B (en) * 2018-11-20 2024-05-10 中山市远尚光电科技有限公司 Automatic ranging-based tele lens
CN115379108A (en) * 2018-11-20 2022-11-22 中山市远尚光电科技有限公司 Telephoto lens based on automatic distance measurement
CN109831609A (en) * 2019-03-05 2019-05-31 上海炬佑智能科技有限公司 TOF depth camera and its Atomatic focusing method
CN110262055A (en) * 2019-05-24 2019-09-20 深圳奥比中光科技有限公司 Projective module group, depth camera and the method for adjusting projective module group focal length
CN112749610A (en) * 2020-07-27 2021-05-04 腾讯科技(深圳)有限公司 Depth image, reference structured light image generation method and device and electronic equipment
CN112328150A (en) * 2020-11-18 2021-02-05 贝壳技术有限公司 Automatic screenshot method, device and equipment, and storage medium
CN112328150B (en) * 2020-11-18 2022-03-29 贝壳技术有限公司 Automatic screenshot method, device and equipment, and storage medium
CN113382165A (en) * 2021-05-31 2021-09-10 上海集成电路制造创新中心有限公司 Camera focusing method and camera focusing system
CN113466339A (en) * 2021-06-18 2021-10-01 之江实验室 Ultrasonic scanning microscope global focusing method and device combined with depth camera
CN113466339B (en) * 2021-06-18 2024-02-09 之江实验室 Global focusing method and device of ultrasonic scanning microscope combined with depth camera
CN115037880A (en) * 2022-07-13 2022-09-09 山西工程职业学院 Quick focusing method for airborne camera

Also Published As

Publication number Publication date
CN106454090B (en) 2019-04-09

Similar Documents

Publication Publication Date Title
CN106454090B (en) Atomatic focusing method and system based on depth camera
CN106412433B (en) Atomatic focusing method and system based on RGB-IR depth camera
CA3040006C (en) Device and method for obtaining distance information from views
US9686461B2 (en) Image capturing device and automatic focusing method thereof
US8830357B2 (en) Image processing device and image processing method including a blurring process
US8885091B2 (en) Imaging device and distance information detecting method
CN105578019B (en) Image extraction system capable of obtaining depth information and focusing method
US8830381B2 (en) Imaging device and method to provide bokeh effect in captured image by determining distance and focus of captured objects in secondary image sequence
JP5762211B2 (en) Image processing apparatus, image processing method, and program
CN107533753A (en) Image processing apparatus
US20170324950A1 (en) Methods and apparatus for controlling light field capture
CN105590939B (en) Imaging sensor and output method, phase focusing method, imaging device and terminal
US10249051B2 (en) Depth extracting camera system using multi focus image and operation method thereof
US8810665B2 (en) Imaging device and method to detect distance information for blocks in secondary images by changing block size
JP7378219B2 (en) Imaging device, image processing device, control method, and program
CN109453517A (en) Virtual role control method and device, storage medium, mobile terminal
US20190355101A1 (en) Image refocusing
CN104144284A (en) Imaging apparatus and imaging method
CN108805921A (en) Image-taking system and method
JP5900017B2 (en) Depth estimation apparatus, reconstructed image generation apparatus, depth estimation method, reconstructed image generation method, and program
KR101994473B1 (en) Method, apparatus and program sotred in recording medium for refocucing of planar image
US9094581B2 (en) Imaging device and distance information detecting method
JP2020194454A (en) Image processing device and image processing method, program, and storage medium
KR102298047B1 (en) Method of recording digital contents and generating 3D images and apparatus using the same
CN111258166B (en) Camera module, periscopic camera module, image acquisition method and working method

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant