CN109035320A - Depth extraction method based on monocular vision - Google Patents

Depth extraction method based on monocular vision Download PDF

Info

Publication number
CN109035320A
CN109035320A CN201810913286.0A CN201810913286A CN109035320A CN 109035320 A CN109035320 A CN 109035320A CN 201810913286 A CN201810913286 A CN 201810913286A CN 109035320 A CN109035320 A CN 109035320A
Authority
CN
China
Prior art keywords
camera
image
value
angle
distortion
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201810913286.0A
Other languages
Chinese (zh)
Other versions
CN109035320B (en
Inventor
徐爱俊
武新梅
周素茵
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang A&F University ZAFU
Original Assignee
Zhejiang A&F University ZAFU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang A&F University ZAFU filed Critical Zhejiang A&F University ZAFU
Priority to CN201810913286.0A priority Critical patent/CN109035320B/en
Publication of CN109035320A publication Critical patent/CN109035320A/en
Application granted granted Critical
Publication of CN109035320B publication Critical patent/CN109035320B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Length Measuring Devices By Optical Means (AREA)
  • Studio Devices (AREA)

Abstract

The invention discloses a kind of depth extraction methods based on monocular vision, and step 1: demarcating mobile phone camera, obtain camera internal parameter and image resolution ratio;Step 2: establishing depth extraction model,Step 3: by the Image Acquisition to target to be measured, target point pixel value u, v are obtained;Step 4: the camera internal parameter and target point pixel value and combining camera depth extraction model obtained using above-mentioned steps calculates target to be measured depth value on image.Depth extraction method based on monocular vision of the invention can be suitable for the different cameras of parameters such as field angle, focal length, image resolution ratio, improve range accuracy, provide support for object measurement and real scene three-dimensional reconstruction in machine vision.

Description

Depth extraction method based on monocular vision
Technical field
The present invention relates to ground close-range photogrammetry field, especially a kind of depth of single camera vision system lower pinhole camera Extracting method.
Background technique
Object ranging and positioning based on image, are broadly divided into two methods of initiative range measurement and passive ranging[1].Actively Ranging is to install laser ranging system on machine (such as camera) to carry out ranging[2-4].Compared with initiative range measurement, due to passive ranging Installation laser ranging system is not needed, thus by favor and to be widely applied.
Passive ranging application wider is monocular vision ranging[7-9].The depth information acquisition method of early stage is mainly binocular Stereoscopic vision and camera motion information need multiple image to complete the acquisition of image depth information[10-16], with binocular distance measurement It compares, the acquisition of monocular range images does not need stringent hardware condition, more competitive superiority.In the prior art, monocular vision There are many methods for the object Depth Information Acquistion of system:
The depth information of target to be measured is such as obtained using corresponding points standardization[17-19].Document [17] has studied one kind Robot target positioning distance measuring method based on monocular vision, this method are usually to obtain the inside and outside ginseng of camera by camera calibration Number, solves the transformational relation between image coordinate system and world coordinate system in conjunction with projection model, to calculate object depth Information.Unfortunately, the method needs to acquire the target image of different direction, and accurately record each point in world coordinate system and Respective coordinates in image coordinate system, stated accuracy are affected for measurement accuracy.Document [20] puts object of reference on road surface And its distance is measured, suitable mathematical model is selected, the corresponding relationship being fitted between object of reference distance and pixel recycles this pass It is extract real-time depth information.Unfortunately, the method precision of document [20] will receive telemeasurement error and error of fitting Influence.
Document [21] devises a kind of vertical target image, and the angle point data by detecting the image establish image ordinate Mapping relations between pixel value and actual measurement angle combine known vehicle-mounted monocular camera height to obtain figure using this relationship The vehicle-mounted depth information as in.Since different cameral equipment inner parameter has differences, for the camera apparatus of different model, the party Method needs to resurvey target image information, establishes camera depth and extracts model, and different in-vehicle camera due to camera lens production with The reasons such as assembly, so that camera pitch angle can also have differences, therefore the method versatility of document [21] is poor.In addition, document [21] method using relationship between vertical target research perpendicular picture point imaging angle and ordinate pixel value, and by this Applied to the measurement of Object Depth on horizontal plane, so that range accuracy is relatively low, because camera level and vertical direction distort Rule is not exactly the same.Application No. is 201710849961.3 patent applications, disclose a kind of improved suitable for intelligent sliding The camera calibration model and distortion correction model of moved end camera are (hereinafter referred to as: the improved calibration mold with nonlinear distortion variable Type), this method can help to correct scaling board picture, obtain the inside and outside parameter of camera of higher precision, unfortunately, this method It does not expand in the nonlinear distortion correction and the measurement of object to testing image.
Bibliography:
[1] He Ruofei, Tian Xuetao, Liu Hongjuan wait unmanned plane target localization method of the based on Monte Carlo Kalman filtering [J] Northwestern Polytechnical University journal, 2017,35 (3): 435-441.
[2] Lin F, Dong X, Chen B M, et al.A Robust Real-Time Embedded Vision System on an Unmanned Rotorcraft for Ground Target Following[J].IEEE Trans on Industrial Electronics, 2012,59 (2): 1038-1049.
[3] Zhang Wanlin, Hu Zhengliang, Zhu Jianjun wait one of the comprehensive view instrument of individual soldier target position calculation method [J] Electronic measurement technique, 2014,37 (11): 1-3.
[4] Sun Junling, Sun Guangmin, Ma Pengge wait laser eyepiece of the based on symmetrical wavelet noise reduction and asymmetric Gauss curve fitting Position [J] Chinese laser, 2017,44 (6): 178-185.
[5] Shi Jie, Li Yin child, Qi Guoqing, passive tracking algorithm [J] China under waiting not exclusively to measure based on machine vision Middle University of Science and Technology's journal, 2017,45 (6): 33-37.
[6] Xu Cheng, yellow grand celebration, a kind of passive target positioning of the numerous small drone of clanging or clanking sound in hole and precision analysis [J] instrument instrument Table journal, 2015,36 (5): 1115-1122.
[7] Li Kehong, Jiang Lingmin, Gong Yong justice .2 are tieed up to 3 d image/Video Quality Metric image depth extracting method and are summarized [J] Journal of Image and Graphics, 2014,19 (10): 1393-1406.
[8] Wang Hao, Xu Zhiwen, Xie Kun, wait binocular range-measurement system [J] Jilin University journal of the based on OpenCV, and 2014, 32 (2): 188-194.
[9] Sun W, Chen L, Hu B, et al.Binocular vision-based position determination algorithm and system[C]//Proceedings of the 2012International Conference on Computer Distributed Control and Intelligent Environmental Monitoring.Piscataway:IEEE Computer Society, 2012:170-173.
[10]Ikeuchi K.Determining a depth map using a dual photometric stereo [J] .The International Journal of Robotics Research, 1987,6 (1): 15-31.
[11] Shao M, Simchony T, Chellappa R.New algorithms from reconstruction of a3-d depth map from one or more images[C]//Proceedings of CVPR’88.Ann Arbor:IEEE, 1988:530-535.
[12] Matthies L, Kanade T, Szeliski R.Kalman filter-based algorithms for estimating depth from image sequences[J].International Journal ofComputer Vision, 1989,3 (3): 209-238.
[13] Mathies L, Szeliski R, Kanade T.Incremental estimation of dense Depth maps from image sequence [C] //Proceedings of CVPR ' 88.Ann Arbor:IEEE, 1988:366-374.
[14] Mori T, Yamamoto M.A dynamic depth extraction method [C] // Proceedings of Third International Conference on Computer Vision.Osaka:IEEE, 1990:672-676.
[15] Inoue H, Tachikawa T, Inaba M.Robot vision system with a Correlation chip for real-time tracking, optical flow and depth map generation [C] //Proceeding of Robotics and Automation.Nice:IEEE, 1992:1621-1626.
[16] Tree image distance measuring method [J] the agricultural mechanics of Hu Tianxiang, Zheng Jiaqiang, Zhou Hongping based on binocular vision Report, 2010,41 (11): 158-162.
[17] Yu Naigong, Huang Can, Lin Jia are calculated based on robot target positioning distance measuring technique study [J] of monocular vision Machine measurement and control, 2012,20 (10): 2654-2660.
[18] ranging research [J] the robot in Wu Gang, Tang Zhen people's monocular formula autonomous robot vision guided navigation, 2010, 32 (6): 828-832.
[19] Lu Weiwei, Xiao Zhitao, Lei Meilin study [J] with distance measuring method based on the front vehicles detection of monocular vision Video Applications and engineering, 2011,35 (1): 125-128.
[20] Wu C F, Lin C J, Lee C Y, et al.Applying a functional neurofuzzy network to real-time lane detection and front-vehicle distance measurement [J] .IEEE Transactions on Systems, Man and Cybemetics-Part C:Applications and Reviews, 2012,42 (4): 577-589.
[21] yellow cloudling, peak, Xu Guoyan wait monocular depth information extraction [J] of based on single width vertical target image BJ University of Aeronautics & Astronautics's journal, 2015,41 (4): 649-655.
Summary of the invention
The object of the present invention is to provide a kind of depth extraction methods based on monocular vision, can be suitable for field angle, coke Away from different cameras of parameters such as, image resolution ratios, measurement accuracy is improved, is object dimensional measurement and true field in machine vision Scape three-dimensional reconstruction provides support.
To achieve the above object, the present invention adopts the following technical scheme:
A kind of depth extraction method based on monocular vision, it is characterised in that include the following steps:
Step 1: demarcating mobile phone camera, obtains camera internal parameter and image resolution ratio
Using Zhang Zhengyou calibration method, and the improved peg model with nonlinear distortion variable is introduced to camera internal parameter It is corrected
The physical size size as pixel each in plane is set first as dx*dy, and image coordinate system (x, y) origin is in picture Coordinate in plain coordinate system (u, v) is (u0, v0), (x, y) is the normalized coordinate of picture point in real image, any picture in image Element meets following relationship in two coordinate systems:
fx、fyFor the normalization focal length in x-axis and y-axis, any point P in camera coordinates systemc(Xc, Yc, Zc) project to image It is (x on coordinate systemc, yc, f), it is f with initial point distance that image coordinate system plane is vertical with optical axis z-axis, according to similar triangles original Reason it follows that
Introduce the improved peg model with nonlinear distortion variable, including the diameter as caused by lens shape defect To distortion and since there are tangential distortion caused by different degrees of bias, radial distortion mathematical models for optical system are as follows:
Wherein r2=x2+y2, (x ', y ') it is sat for the normalization of the ideal linearity camera coordinates system without distortion term after correction Scale value, radial distortion value is related with the position of picture point in the picture, and the radial distortion value at image border is larger,
Tangential distortion mathematical model are as follows:
It wherein include k1、k2、k3、p1、p2Totally 5 kilrrfactors obtain distortion correction Function Modules by formula (3), (4) Type is as follows:
There are following relationships for conversion from world coordinate transformation to camera coordinates:
Pc=R (PW- C)=RPW+T (6)
Convolution (1)~(6), may be expressed as: with homogeneous coordinates and matrix form
Mint、MextIt is the inside and outside parameter matrix of camera calibration respectively, wherein camera internal parameter includes image center pixel Value u0、v0, fx、fyFor in x-axis and y-axis normalization focal length, by Java combination OpenCV realize mobile phone camera calibration, acquisition Inner parameter described in mobile phone camera, camera lens distortion parameter and image resolution ratio vmax、umax
Step 2: depth extraction model is established
Abstract function is set according to the linear relationship between object imaging angle α and ordinate pixel value v, establishes and contains mesh Mark tri- object imaging angle α, ordinate pixel value v and camera rotation angle β parameter space relational models, i.e. α=F (v, β),
Under equipment and camera the rotation angle of different model, subject ordinate pixel value and imaging angle are in pole Significant negative linear correlation, and the slope of the linear relationship and intercept are different, therefore set:
α=F (v, β)=av+b (17)
Wherein parameter a, b is related with camera model and camera rotation angle,
When α is minimized α=αminWhen=90- θ-β, θ is the half at camera vertical field of view angle, i.e. subject projects When to picture lowermost end, v=vmax(vmaxFor camera CMOS or ccd image sensor column coordinate valid pixel number), substitute into formula (17) it can obtain:
90- β-θ=avmax+b (18)
Work as αminWhen+2 90 ° of θ >, i.e. θ > β, camera upward angle of visibility is higher than horizontal line at this time, and ground level unlimited distance, α is unlimited Close to 90 °, v is substantially equal to v at this time0-tanβ*fy, fyFor the focal length of camera under pixel unit, β is negative value, that is, camera inverse time When needle rotates also similarly, therefore, substituting into formula (17) can obtain:
90=a (v0-tanβ·fy)+b (19)
Work as αminWhen+2 90 ° of θ <, i.e. θ < β, camera upward angle of visibility is lower than horizontal line, ground level unlimited distance object at this time Imaging angle α is maximized, αmaxminWhen+2 θ=90- β+θ, i.e., when subject projects to picture highest point, v=0, Substitution formula (17) can obtain:
90- β+θ=b (20)
According to pinhole camera aufbauprinciple, the tangent value of the camera vertical field of view angle θ of half is schemed equal to camera CMOS or CCD As the half of sensor side length is divided by camera focus, therefore θ can be calculated:
L in formula (21)CMOSFor the side length of camera CMOS or ccd image sensor, convolution (18)~(21), F (α, β) Are as follows:
δ is camera nonlinear distortion variable error in formula (10);In conjunction with mobile phone camera shooting height h, according to trigonometric function Principle establishes mobile phone camera depth extraction model:
Step 3: by the Image Acquisition to target to be measured, target point pixel value u, v are obtained;
It further include the nonlinear distortion correction to target to be measured image in the image acquisition step to target to be measured And pretreatment, it may be assumed that
Image Acquisition is carried out by mobile phone camera, establishes perspective geometry model, wherein f is camera focus, and θ is that camera is vertical The half of field angle, h are that camera is taken pictures highly, and β is rotation angle of the camera along camera coordinates system ox axis, and camera rotates clockwise β Value is positive, and is negative counterclockwise, and β value is obtained by camera internal gravity sensor, and α is object imaging angle;
In conjunction with the camera lens distortion parameter that step 1 camera calibration obtains, to radial distortion existing for image and tangential abnormal Become error and carries out nonlinear distortion correction;Ideal linearity normalized coordinate value (x, y) after correction is substituted into formula (1), asks calculation Image each point pixel coordinate value after correcting out, by the method for bilinear interpolation to pixel value after correction carry out interpolation processing to Image after being corrected;The image after correction is pre-processed using computer vision and image processing techniques, including image Binaryzation, morphological image operation and the detection of object contour edge, obtain the edge of object, and then calculate object and ground The geometric center point pixel value (u, v) at the edge of face contact;
Step 4: the camera internal parameter and target point pixel value and combining camera depth extraction obtained using above-mentioned steps Model calculates target to be measured depth value D on image
The size relation between angle beta and the camera vertical field of view angle θ of half is rotated according to camera, selects corresponding depth Model is extracted, step 1 is asked to the camera internal parametric image central point pixel value v of calculation0, normalized focal length f in y-axisyAnd Image resolution ratio vmaxAnd step 3 asks the target to be measured ordinate pixel value v of calculation, camera rotation angle beta and mobile phone camera to clap It takes the photograph height h and substitutes into the depth extraction model, calculate target point depth value D.
Compared with prior art, the beneficial effects of the present invention are: due to the adoption of the above technical scheme,
(1) compared with prior art, method of the invention does not need large scene calibration place, avoids data fitting and causes Error;
(2) there is equipment interoperability using the depth extraction model that the method for the present invention is established, and camera rotation angle is introduced Model, for the camera of different model, it is only necessary to after obtaining camera internal parameter by camera calibration for the first time, single width can be calculated The depth of any picture point on picture.Verified, when distance is to carry out short distance ranging in 0.5~2.6m, depth value measurement is flat Equal relative error is 0.937%, and when distance is 3~10m, measurement relative error is 1.71%.Therefore, it is extracted using this method Depth, measurement accuracy with higher.
Detailed description of the invention
Fig. 1 is distance measuring method flow chart of the invention;
Fig. 2 is novel target figure;
Fig. 3 is Corner Detection Algorithm implementation flow chart;
Fig. 4 is that camera upward angle of visibility is higher than horizontal line shooting geometrical model figure;
Fig. 5 is camera upward angle of visibility lower than horizontal line shooting geometrical model figure;
Fig. 6 is camera shooting perspective geometry model;
Fig. 7 is the relational graph between three kinds of model device object ordinate pixel values and imaging angle;
Fig. 8 is the relational graph between different cameral rotation angle object ordinate pixel value and actual imaging angle.
Specific embodiment
In order to be more clear technical solution of the present invention, below in conjunction with attached drawing 1 to 8, the present invention is described in detail. It should be understood that specific embodiment described in this specification is not intended to limit just for the sake of explaining the present invention Protection scope of the present invention.
The present invention is a kind of depth extraction method based on monocular vision, is included the following steps:
One, mobile phone camera is demarcated, obtains camera internal parameter and image resolution ratio.The calibration uses Zhang Zhengyou Standardization, and introduce the improved peg model with nonlinear distortion variable and camera internal parameter is corrected.
The physical size size as pixel each in plane is set first as dx*dy (unit: mm), image coordinate system (x, Y) coordinate of the origin in pixel coordinate system (u, v) is (u0, v0), (x, y) is the normalized coordinate of picture point in real image, figure Any pixel meets following relationship in two coordinate systems as in:
fx、fyFor the normalization focal length in x-axis and y-axis, any point P in camera coordinates systemc(Xc, Yc, Zc) project to image It is (x on coordinate systemc, yc, f), it is f with initial point distance that image coordinate system plane is vertical with optical axis z-axis, according to similar triangles original Reason it follows that
Introduce the improved peg model with nonlinear distortion variable, including the diameter as caused by lens shape defect To distortion and since there are tangential distortion caused by different degrees of bias, radial distortion mathematical models for optical system are as follows:
Wherein r2=x2+y2, (x ', y ') it is sat for the normalization of the ideal linearity camera coordinates system without distortion term after correction Scale value, radial distortion value is related with the position of picture point in the picture, and the radial distortion value at image border is larger,
Tangential distortion model mathematical model are as follows:
It wherein include k1、k2、k3、p1、p2Totally 5 kilrrfactors obtain distortion correction Function Modules by formula (3), (4) Type is as follows:
There are following relationships for conversion from world coordinate transformation to camera coordinates:
Pc=R (PW- C)=RPW+T (6)
Convolution (1)~(6), may be expressed as: with homogeneous coordinates and matrix form
Mint、MextIt is the inside and outside parameter matrix of camera calibration respectively, wherein camera internal parameter includes image center pixel Value u0、v0, fx、fyFor in x-axis and y-axis normalization focal length, by Java combination OpenCV realize mobile phone camera calibration, acquisition Inner parameter described in mobile phone camera, camera lens distortion parameter and image resolution ratio vmax、umax
Two, it by the acquisition to novel target image, establishes camera depth and extracts model.Existing target is that length and width are equal Black and white chessboard case marker target.The difference of modulation of the invention and existing target is, setting target is apart from camera nearest One row's grid size is d*d mm, and the width of subsequent every row's grid is to fix, and the previous row's value added of the latter parallelism of length is
X in following formulaiFor the actual range of i-th of angle point to camera, yiFor the length of each grid, then adjacent square length Difference DELTA diAre as follows:
If the relationship between the computational length and actual range of each grid is f (x), can be obtained according to formula (8):
Through Pearson correlation analysis, it is between length and actual range extremely significant linear relationship (p < 0.01), Correlation coefficient r be equal to 0.975, by least square method can in the hope of calculate f (x) derivative f ' (x),
Therefore, when a target row grid size nearest apart from camera is d*d mm (the survey when range of d takes 30~60mm Accuracy of measurement highest) when, subsequent every row's width fixes, length incrementFor d*f ' (x) mm, novel target as shown in Fig. 2,
There are the angles that perspective transform phenomenon makes Harris and Shi-Tomasi etc. common when object on shooting level ground Point detection algorithm robustness is poor, and can also detect mistake when camera is larger along camera coordinates system ox axis rotated counterclockwise by angle It loses, therefore the checkerboard angle point detection process based on growth of the propositions such as present invention combination Andreas Geiger and OpenCV are mentioned The comerSubPix () function of confession carries out the detection of sub-pixel corner location, and the algorithm robustness is high, larger to distortion degree Picture extraction effect it is preferable,
The implementation process of Corner Detection Algorithm as shown in figure 3, the above-mentioned modulation of the present invention sub-pixel angle point grid Step are as follows:
1) angle point is found according to the similarity parameter of pixel each in image and template on the image, positions target angle point position It sets;
Two different angle point templates are defined first, and a kind of for the angle point parallel with reference axis, another kind is for rotating 45 ° of angle point, each template are made of 4 filtering cores { A, B, C, E }, with carrying out convolution operation with image later;Then sharp The similarity of each inflection point and angle point is calculated with the two angle point templates:
WhereinIndicate that convolution kernel X (X=A, B, C, E) and template i (i=1,2) are responded in the convolution of some pixel, WithIt indicates the similarity of two kinds of possible inflection points of template i, calculates the available angle of similarity of each pixel in image Point similar diagram;It is handled using non-maxima suppression algorithm angle steel joint pixel map to obtain candidate point;Then it is counted with gradient Method verify these candidate points in the n x n neighborhood of a local, first local area grayscale image carries out sobel filtering, then Weighting direction histogram (32bins) is calculated, finds two therein main mode γ with meanshift algorithm1And γ2;Root According to the direction at edge, for desired gradient intensityConstruct a template T.(* indicates cross-correlation operation Symbol) and angle point similarity product as angle point score value, then judged just to obtain initial angle point with threshold value.
2) the position and direction progress sub-pixel of angle steel joint finely extracts;
Sub-pixel Corner character is carried out with the comerSubPix () function in OpenCV, by Corner character to sub- picture Element, to obtain the other Corner Detection effect of sub-pixel;To refine edge direction vector, it is minimized according to image gradient value Standard deviation rate:
WhereinIt is adjacent pixel collection, the gradient value m with module if= [cos(γi)sin(γi)]TMatch.(ask calculation scheme according to document Geiger A, Moosmann F, Caret al.Automatic camera and range sensor calibration using a single shot[C]// Robotics and Automation (ICRA), 2012 IEEE International Conference on.IEEE, 2012:3936-3943.)
3) it is finally label angle point and exports its subpixel coordinate, gridiron pattern is grown and rebuild according to energy function, marks Remember angle point, exports sub-pixel angular coordinate;
According to document " Geiger A, Moosmann F, Caret al.Automatic camera and range Sensor calibration using a single shot [C] //Robotics and Automation (ICRA), 2012IEEE International Conference on.IEEE, the method that 2012:3936-3943. " is provided optimize energy Function rebuilds gridiron pattern and marks angle point, energy growth function formula are as follows:
E (x, y)=Ecorners(y)+Estruct(x, y) (16)
Wherein, EcornersIt is the negative value of current chessboard angle point sum, EstructIt is of two adjacent corner points and prediction angle point With degree;Angle point pixel value is exported by OpenCV.
Linear correlative analysis is carried out to image objects angle, ordinate pixel value using SPSS 22, exports Pearson phase Relationship number, verified, under equipment and camera the rotation angle of different model, object ordinate pixel value is in actual imaging angle Extremely significant negative correlativing relation (p < 0.01), in addition, the present invention is also vertical to object under different device models and camera rotation angle The slope difference of linear function carries out significance test between coordinate pixel value and imaging angle, the results showed that, distinct device type Number and camera rotation angle under between object ordinate pixel value and imaging angle linear function heteropolar significant (the p < of slope differences 0.01), illustrating the equipment and camera rotation angle of different model, depth extraction model is different,
Abstract function is set according to the linear relationship between object imaging angle α and ordinate pixel value v, establishes and contains mesh Mark tri- object imaging angle α, ordinate pixel value v and camera rotation angle β parameter space relational models, i.e. α=F (v, β),
Under equipment and camera the rotation angle of different model, subject ordinate pixel value and imaging angle are in pole Significant negative linear correlation, and the slope of the linear relationship and intercept are different, therefore set:
α=F (v, β)=av+b (17)
Wherein parameter a, b is related with camera model and camera rotation angle,
When α is minimized α=αminWhen=90- θ-β, θ is the half at camera vertical field of view angle, i.e. subject projects When to picture lowermost end, v=vmax(vmaxFor camera CMOS or ccd image sensor column coordinate valid pixel number), substitute into formula (17) it can obtain:
90- β-θ=avmax+b (18)
Work as αminWhen+2 90 ° of θ >, i.e. θ > β, camera upward angle of visibility is higher than horizontal line at this time, and camera shoots perspective geometry model Such as Fig. 4, ground level unlimited distance, α is infinitely close to 90 °, and v is substantially equal to v at this time0-tanβ*fy, fyFor phase under pixel unit The focal length of machine, when β rotates counterclockwise for negative value, that is, camera also similarly, therefore, substituting into formula (17) can obtain:
90=a (v0-tanβ·fy)+b (19)
Work as αminWhen+2 90 ° of θ <, i.e. θ < β, camera upward angle of visibility is lower than horizontal line at this time, and camera shoots perspective geometry model If Fig. 5, ground level unlimited distance object imaging angle α are maximized, αmaxminWhen+2 θ=90- β+θ, i.e. subject When body projects to picture highest point, v=0, substituting into formula (17) can be obtained:
90- β+θ=b (20)
According to pinhole camera aufbauprinciple, the tangent value of the camera vertical field of view angle θ of half is schemed equal to camera CMOS or CCD As the half of sensor side length is divided by camera focus, therefore θ can be calculated:
L in formula (21)CMOSFor the side length of camera CMOS or ccd image sensor, convolution (18)~(21), F (α, β) Are as follows:
δ is camera nonlinear distortion variable error in formula (10), in conjunction with mobile phone camera shooting height h, according to trigonometric function Principle establishes mobile phone camera depth extraction model:
Three, by the Image Acquisition to target to be measured, target point pixel value u, v are obtained.Figure is carried out by mobile phone camera As acquisition, perspective geometry model such as Fig. 6 is established, wherein f is camera focus, and θ is the half at camera vertical field of view angle, and h is camera It takes pictures highly, β is rotation angle of the camera along camera coordinates system ox axis, and camera rotates clockwise β value and is positive, is negative counterclockwise, β value It is obtained by camera internal gravity sensor, α is object imaging angle;The camera lens obtained in conjunction with first step camera calibration Distortion parameter carries out nonlinear distortion correction to radial distortion existing for image and tangential distortion error;By the ideal after correction Linear normalization coordinate value (x, y) substitutes into formula (1), image each point pixel coordinate value after asking calculating to correct, by bilinearity Image after slotting method corrects pixel value progress interpolation processing after correction;Using computer vision and image procossing Technology pre-processes the image after correction, including image binaryzation, morphological image operation and the inspection of object contour edge It surveys, obtains the edge of object, and then calculate the geometric center point pixel value (u, v) at the edge of object and ground face contact.
Four, the camera internal parameter and target point pixel value and combining camera depth extraction mould of above-mentioned steps acquisition are utilized Type calculates target to be measured depth value D on image.It is rotated between angle beta and the camera vertical field of view angle θ of half according to camera Size relation, select corresponding depth model, above-mentioned steps asked to the camera internal parametric image central point pixel value v of calculation0, Normalized focal length f in y-axisyAnd image resolution ratio vmaxAnd above-mentioned steps ask calculation target to be measured ordinate pixel value v, Camera rotates angle beta and mobile phone camera shooting height h substitutes into the depth extraction model, calculates target point depth value D.
Embodiment 1
Below by taking millet 3 (MI 3) mobile phone as an example, the depth for illustrating the optimization of the invention based on monocular vision is mentioned It takes and passive ranging method.
One, mobile phone camera is demarcated, obtains camera internal parameter and image resolution ratio
Use ranks number be 8*9 size be 20*20 gridiron pattern scaling board as the experimental material of camera calibration, lead to The scaling board picture that 3 mobile phone camera of millet acquires 20 different angles is crossed, using OpenCV according to above-mentioned improved with non-thread The camera calibration model of sex distortion item demarcates millet 3 (MI 3) mobile phone camera,
Scaling board picture is read using fin () function first, and obtains the image of the first picture by .cols .rows Resolution ratio;Then sub-pixel angle point in scaling board picture is extracted by find4QuadCornerSubpix () function, be used in combination DrawChessboardCorners () function marks angle point;CalibrateCamera () function is called to demarcate camera, It is used for obtained camera interior and exterior parameter to carry out projection again to the three-dimensional point in space calculating, obtains new subpoint, calculate Error between new subpoint and old subpoint;Camera internal reference matrix and distortion parameter are exported and save,
Calibration gained camera internal parameter are as follows: fx=3486.5637, u0=1569.0383, fy=3497.4652, v0= 2107.9899, image resolution ratio is 3120 × 4208, camera lens distortion parameter are as follows: [0.0981, -0.1678,0.0003, - 0.0025,0.0975],
Two, it by the acquisition to novel target image, establishes camera depth and extracts model
The initial experiment material that the present invention uses traditional gridiron pattern scaling board of 45*45mm to design as target, to calculate The difference of adjacent square length, the present invention devise 6 groups of experiments, extract traditional X-comers that grid size is 45*45mm Value, and ask calculate adjacent corner points between the actual physics distance that is represented under world coordinate system of unit pixel, to guarantee to indulge between angle point Coordinate pixel value difference is roughly equal, and the value of the length yi of each grid is as shown in table 1,
The calculating width of each grid of table 1
Table 1 Computing width of each grid
Through Pearson correlation analysis, it is between length and actual range extremely significant linear relationship (p < 0.01), Correlation coefficient r is equal to 0.975, can be in the hope of derivative f ' (x)=0.262 of calculating f (x), therefore, when the mark by least square method When a range row grid size nearest from camera is 45*45mm, then every row's width is fixed, width value added Δ d is 11.79mm,
The angle point of the novel target is extracted by the Robust Algorithm of Image Corner Extraction in specific implementation step,
The present invention choose respectively millet, Huawei, tri- kinds of different models of iPhone smart phone as image capture device, Camera rotates angle beta={ -10 °, 0 °, 10 °, 20 °, 30 ° }.Data are acquired using the Corner Detection Algorithm, and to its relationship Carry out Function Fitting, Fig. 7 be β=10 ° when three kinds of different models smart phone ordinate pixel value and image objects angle it Between relationship, Fig. 8 is that different cameral rotates relationship between ordinate pixel value and image objects angle under angle,
The camera apparatus and camera of different model rotate angle, with the increase of ordinate pixel value, image objects angle Taper off trend, and the difference of device model and camera rotation angle makes that different lines are presented between pixel value and imaging angle Property functional relation, using SPSS 22 to image objects angle, ordinate pixel value carry out Linear correlative analysis, export Pearson Correlation coefficient r is as shown in table 2.
2 object ordinate pixel value of table and imaging angle related coefficient
Table 2 Pearson correlation coefficient of image ordinate pixel values and actual imaging angles
Note: * * indicates extremely significant (p < 0.01).
Note:**represents very significant correlation (p < 0.01)
Verified, equipment and camera with model rotate under angle, and object ordinate pixel value is in actual imaging angle Extremely significant negative correlativing relation (p < 0.01), correlation coefficient r are greater than 0.99.In addition, the present invention is also to different device model and phase The slope difference that machine rotates linear function between object ordinate pixel value and imaging angle under angle carries out significance test.Knot Fruit show distinct device model and camera rotation angle under between object ordinate pixel value and imaging angle linear function it is oblique Rate difference is extremely significant (p < 0.01), illustrates the equipment and camera rotation angle of different model, depth extraction model is not Together.
3 mobile phone camera inner parameter of millet is substituted into formula (10) and obtained by the depth extraction model according to specific embodiment:
The specific depth extraction model of the equipment is obtained according to trigonometric function principle are as follows:
Three, by the Image Acquisition to target to be measured, target point pixel value u, v are obtained.
Use (MI 3) camera of millet mobile phone 3 as picture collection equipment, carries out picture collection by camera trivets, and The height h for measuring camera to ground is equal to 305mm, and camera rotation angle β is equal to 0 °,
Nonlinear distortion correction is carried out to radial distortion existing for image and tangential distortion error;
The camera lens distortion parameter obtained according to first step camera calibration: [0.0981, -0.1678,0.0003, - 0.0025,0.0975], ideal linearity normalized coordinate value after correcting is calculated according to formula (5):
Image each point pixel coordinate value after correcting is calculated in conjunction with formula (1) and (2), is handled and is rectified by bilinear interpolation Image after just;
The present invention measures its depth and distance by taking the cuboid box being placed on level ground as an example, first to acquisition Image carries out binary conversion treatment, then carries out edge detection to cuboid box using Canny operator, extracts object profile. Extracting cuboid box bottom margin central point pixel value is (1851.23,3490).
Four, the camera internal parameter and target point pixel value and combining camera depth extraction mould of above-mentioned steps acquisition are utilized Type, calculate object to be measured object image take up an official post meaning point depth value.
Camera internal parameter, camera are taken pictures height h, rotation angle beta and cuboid box bottom margin central point is vertical sits Mark pixel value v, which substitutes into formula (24), can calculate the object actual imaging angle equal to 69.58 °.According to trigonometric function original Reason calculates target point depth value D (unit: mm):
D=305*tan 69.58 °=819.21 (27)
By tape measuring, the distance of the cuboid box to phase machine side is 825mm, therefore the present invention is used to carry out ranging, Its relative error is 0.70%.

Claims (1)

1. a kind of depth extraction method based on monocular vision, it is characterised in that include the following steps:
Step 1: demarcating mobile phone camera, obtains camera internal parameter and image resolution ratio
Using Zhang Zhengyou calibration method, and introduces the improved peg model with nonlinear distortion variable and camera internal parameter is carried out Correction
The physical size size as pixel each in plane is set first as dx*dy, and image coordinate system (x, y) origin is sat in pixel Coordinate in mark system (u, v) is (u0, v0), (x, y) is the normalized coordinate of picture point in real image, and any pixel exists in image Meet following relationship in two coordinate systems:
fx、fyFor the normalization focal length in x-axis and y-axis, any point P in camera coordinates systemc(Xc, Yc, Zc) project to image coordinate It fastens as (xc, yc, f), it is f with initial point distance that image coordinate system plane is vertical with optical axis z-axis, can according to similar triangle theory To obtain:
The improved peg model with nonlinear distortion variable is introduced, including the radial direction as caused by lens shape defect is abnormal Become and since there are tangential distortion caused by different degrees of bias, radial distortion mathematical models for optical system are as follows:
Wherein r2=x2+y2, (x ', y ') it is the normalized coordinate value that the ideal linearity camera coordinates system of distortion term is free of after correcting, Radial distortion value is related with the position of picture point in the picture, and the radial distortion value at image border is larger,
Tangential distortion mathematical model are as follows:
It wherein include k1、k2、k3、p1、p2Totally 5 kilrrfactors obtain distortion correction function model such as by formula (3), (4) Under:
There are following relationships for conversion from world coordinate transformation to camera coordinates:
Pc=R (PW- C)=RPW+T (6)
Convolution (1)~(6), may be expressed as: with homogeneous coordinates and matrix form
Mint、MextIt is the inside and outside parameter matrix of camera calibration respectively, wherein camera internal parameter includes image center pixel value u0、 v0, fx、fyFor in x-axis and y-axis normalization focal length, by Java combination OpenCV realize mobile phone camera calibration, acquisition mobile phone phase Inner parameter described in machine, camera lens distortion parameter and image resolution ratio vmax、umax
Step 2: depth extraction model is established
Abstract function is set according to the linear relationship between object imaging angle α and ordinate pixel value v, establishes and contains object Imaging angle α, ordinate pixel value v and tri- parameter space relational models of camera rotation angle β, i.e. α=F (v, β),
Under equipment and camera the rotation angle of different model, subject ordinate pixel value is in extremely significant with imaging angle Negative linear correlation, and the slope of the linear relationship and intercept are different, therefore set:
α=F (v, β)=av+b (17)
Wherein parameter a, b is related with camera model and camera rotation angle,
When α is minimized α=αminWhen=90- θ-β, θ is the half at camera vertical field of view angle, i.e. subject projects to figure When piece lowermost end, v=vmax(vmaxFor camera CMOS or ccd image sensor column coordinate valid pixel number), substituting into formula (17) can :
90- β-θ=avmax+b (18)
Work as αminWhen+2 90 ° of θ >, i.e. θ > β, camera upward angle of visibility is higher than horizontal line, ground level unlimited distance, α infinite approach at this time In 90 °, v is substantially equal to v at this time0-tanβ*fy, fyFor the focal length of camera under pixel unit, β is that negative value, that is, camera revolves counterclockwise When turning also similarly, therefore, substituting into formula (17) can obtain:
90=a (v0-tanβ·fy)+b (19)
Work as αminWhen+2 90 ° of θ <, i.e. θ < β, camera upward angle of visibility is lower than horizontal line, the imaging of ground level unlimited distance object at this time Angle [alpha] is maximized, αmaxminWhen+2 θ=90- β+θ, i.e., when subject projects to picture highest point, v=0 is substituted into Formula (17) can obtain:
90- β+θ=b (20)
According to pinhole camera aufbauprinciple, the tangent value of the camera vertical field of view angle θ of half is equal to camera CMOS or ccd image passes The half of sensor side length can calculate θ divided by camera focus:
L in formula (21)CMOSFor the side length of camera CMOS or ccd image sensor, convolution (18)~(21), F (α, β) are as follows:
δ is camera nonlinear distortion variable error in formula (10);In conjunction with mobile phone camera shooting height h, according to trigonometric function principle Establish mobile phone camera depth extraction model:
Step 3: by the Image Acquisition to target to be measured, target point pixel value u, v are obtained;
It further include to the nonlinear distortion of target to be measured image correction and pre- in the image acquisition step to target to be measured Processing, it may be assumed that
Image Acquisition is carried out by mobile phone camera, establishes perspective geometry model, wherein f is camera focus, and θ is camera vertical field of view The half at angle, h are that camera is taken pictures highly, and β is rotation angle of the camera along camera coordinates system ox axis, and camera rotates clockwise β value and is Just, it is negative counterclockwise, β value is obtained by camera internal gravity sensor, and α is object imaging angle;
In conjunction with the camera lens distortion parameter that step 1 camera calibration obtains, radial distortion existing for image and tangential distortion are missed Difference carries out nonlinear distortion correction;Ideal linearity normalized coordinate value (x, y) after correction is substituted into formula (1), calculating is asked to rectify Image each point pixel coordinate value after just carries out interpolation processing to pixel value after correction by the method for bilinear interpolation to obtain Image after correction;The image after correction is pre-processed using computer vision and image processing techniques, including image two-value Change, morphological image operation and the detection of object contour edge, obtain the edge of object, and then calculate object and connect with ground The geometric center point pixel value (u, v) at the edge of touching;
Step 4: the camera internal parameter and target point pixel value and combining camera depth extraction mould obtained using above-mentioned steps Type calculates target to be measured depth value D on image
The size relation between angle beta and the camera vertical field of view angle θ of half is rotated according to camera, selects corresponding depth extraction Step 1 is sought the camera internal parametric image central point pixel value v of calculation by model0, normalized focal length f in y-axisyAnd image Resolution ratio vmaxAnd step 3 asks the target to be measured ordinate pixel value v of calculation, camera rotation angle beta and mobile phone camera shooting high It spends h and substitutes into the depth extraction model, calculate target point depth value D.
CN201810913286.0A 2018-08-12 2018-08-12 Monocular vision-based depth extraction method Active CN109035320B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810913286.0A CN109035320B (en) 2018-08-12 2018-08-12 Monocular vision-based depth extraction method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810913286.0A CN109035320B (en) 2018-08-12 2018-08-12 Monocular vision-based depth extraction method

Publications (2)

Publication Number Publication Date
CN109035320A true CN109035320A (en) 2018-12-18
CN109035320B CN109035320B (en) 2021-08-10

Family

ID=64633772

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810913286.0A Active CN109035320B (en) 2018-08-12 2018-08-12 Monocular vision-based depth extraction method

Country Status (1)

Country Link
CN (1) CN109035320B (en)

Cited By (38)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109509200A (en) * 2018-12-26 2019-03-22 深圳市繁维医疗科技有限公司 Checkerboard angle point detection process, device and computer readable storage medium based on contours extract
CN109708655A (en) * 2018-12-29 2019-05-03 百度在线网络技术(北京)有限公司 Air navigation aid, device, vehicle and computer readable storage medium
CN109730683A (en) * 2018-12-21 2019-05-10 重庆金山医疗器械有限公司 Endoscope object size calculation method and analysis system
CN109883654A (en) * 2019-01-25 2019-06-14 武汉精立电子技术有限公司 A kind of chessboard trrellis diagram, generation method and localization method for OLED sub-pixel positioning
CN109901594A (en) * 2019-04-11 2019-06-18 清华大学深圳研究生院 A kind of localization method and system of weed-eradicating robot
CN110009682A (en) * 2019-03-29 2019-07-12 北京理工大学 A kind of object recognition and detection method based on monocular vision
CN110297498A (en) * 2019-06-13 2019-10-01 暨南大学 A kind of rail polling method and system based on wireless charging unmanned plane
CN110490826A (en) * 2019-08-26 2019-11-22 河南农业大学 Droplet sample image processing method in a kind of camera
CN110517325A (en) * 2019-08-29 2019-11-29 的卢技术有限公司 The vehicle body surrounding objects localization method and system of a kind of coordinate transform and coordinate transform
CN110737942A (en) * 2019-10-12 2020-01-31 清华四川能源互联网研究院 Underwater building model establishing method, device, equipment and storage medium
CN110796709A (en) * 2019-10-29 2020-02-14 上海眼控科技股份有限公司 Method and device for acquiring size of frame number, computer equipment and storage medium
CN110816522A (en) * 2019-11-12 2020-02-21 深圳创维数字技术有限公司 Vehicle attitude control method, apparatus, and computer-readable storage medium
CN110853080A (en) * 2019-09-30 2020-02-28 广西慧云信息技术有限公司 Method for measuring size of field fruit
CN111105488A (en) * 2019-12-20 2020-05-05 成都纵横自动化技术股份有限公司 Imaging simulation method and device, electronic equipment and storage medium
CN111191646A (en) * 2019-12-23 2020-05-22 福建亿榕信息技术有限公司 Intelligent identification method for pointer instrument
CN111192235A (en) * 2019-12-05 2020-05-22 中国地质大学(武汉) Image measuring method based on monocular vision model and perspective transformation
CN111380456A (en) * 2018-12-29 2020-07-07 沈阳高精数控智能技术股份有限公司 Quick detection method for ballastless track plate
CN111568222A (en) * 2020-02-29 2020-08-25 佛山市云米电器科技有限公司 Water dispenser control method, water dispenser and computer readable storage medium
CN111623776A (en) * 2020-06-08 2020-09-04 昆山星际舟智能科技有限公司 Method for measuring distance of target by using near infrared vision sensor and gyroscope
CN112197708A (en) * 2020-08-31 2021-01-08 深圳市慧鲤科技有限公司 Measuring method and device, electronic device and storage medium
CN112258455A (en) * 2020-09-28 2021-01-22 上海工程技术大学 Detection method for detecting spatial position of part based on monocular vision
CN112614191A (en) * 2020-12-16 2021-04-06 江苏智库智能科技有限公司 Loading and unloading position detection method, device and system based on binocular depth camera
CN112798812A (en) * 2020-12-30 2021-05-14 中山联合汽车技术有限公司 Target speed measuring method based on monocular vision
CN113222891A (en) * 2021-04-01 2021-08-06 东方电气集团东方锅炉股份有限公司 Line laser-based binocular vision three-dimensional measurement method for rotating object
CN113344998A (en) * 2021-06-25 2021-09-03 北京市商汤科技开发有限公司 Depth detection method and device, computer equipment and storage medium
CN113496528A (en) * 2021-09-07 2021-10-12 湖南众天云科技有限公司 Method and device for calibrating position of visual detection target in fixed traffic roadside scene
CN113538546A (en) * 2021-09-17 2021-10-22 智道网联科技(北京)有限公司 Target detection method, device and equipment for automatic driving
CN113592934A (en) * 2021-06-29 2021-11-02 浙江工业大学 Monocular camera-based target depth and height measuring method and device
CN113793388A (en) * 2021-08-09 2021-12-14 南京工程学院 Stereoscopic vision interpersonal safe distance detection method based on deep learning
CN114022378A (en) * 2021-11-04 2022-02-08 中天智能装备有限公司 Copper strip shielding layer overlapping rate detection method based on vision
CN114155290A (en) * 2021-11-18 2022-03-08 合肥富煌君达高科信息技术有限公司 System and method for large-field-of-view high-speed motion measurement
CN114549666A (en) * 2022-04-26 2022-05-27 杭州蓝芯科技有限公司 AGV-based panoramic image splicing calibration method
CN114897784A (en) * 2022-04-13 2022-08-12 广东工业大学 Monocular egg size assembly line measuring method
CN115026828A (en) * 2022-06-23 2022-09-09 池州市安安新材科技有限公司 Robot arm grabbing control method and system
CN115272110A (en) * 2022-07-21 2022-11-01 四川大学 Projector distortion correction method and device in structured light three-dimensional reconstruction
CN116091608A (en) * 2023-04-11 2023-05-09 深之蓝海洋科技股份有限公司 Positioning method and positioning device for underwater target, underwater equipment and storage medium
CN116402871A (en) * 2023-03-28 2023-07-07 苏州大学 Monocular distance measurement method and system based on scene parallel elements and electronic equipment
CN116758171A (en) * 2023-08-21 2023-09-15 武汉中导光电设备有限公司 Imaging system pose correction method, device, equipment and readable storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104034514A (en) * 2014-06-12 2014-09-10 中国科学院上海技术物理研究所 Large visual field camera nonlinear distortion correction device and method
CN104331896A (en) * 2014-11-21 2015-02-04 天津工业大学 System calibration method based on depth information
US9741118B2 (en) * 2013-03-13 2017-08-22 Fotonation Cayman Limited System and methods for calibration of an array camera

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9741118B2 (en) * 2013-03-13 2017-08-22 Fotonation Cayman Limited System and methods for calibration of an array camera
CN104034514A (en) * 2014-06-12 2014-09-10 中国科学院上海技术物理研究所 Large visual field camera nonlinear distortion correction device and method
CN104331896A (en) * 2014-11-21 2015-02-04 天津工业大学 System calibration method based on depth information

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
J. CARDILLO,ET AL: "《3-D position sensing using a passive monocular vision system》", 《IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE》 *
冯春: "《基于单目视觉的目标识别与定位研究》", 《中国博士学位论文全文数据库 信息科技辑》 *

Cited By (56)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109730683A (en) * 2018-12-21 2019-05-10 重庆金山医疗器械有限公司 Endoscope object size calculation method and analysis system
CN109730683B (en) * 2018-12-21 2021-11-05 重庆金山医疗技术研究院有限公司 Endoscope target size calculation method and analysis system
CN109509200A (en) * 2018-12-26 2019-03-22 深圳市繁维医疗科技有限公司 Checkerboard angle point detection process, device and computer readable storage medium based on contours extract
CN109509200B (en) * 2018-12-26 2023-09-29 深圳市繁维医疗科技有限公司 Checkerboard corner detection method based on contour extraction and computer readable storage medium
CN109708655A (en) * 2018-12-29 2019-05-03 百度在线网络技术(北京)有限公司 Air navigation aid, device, vehicle and computer readable storage medium
CN111380456A (en) * 2018-12-29 2020-07-07 沈阳高精数控智能技术股份有限公司 Quick detection method for ballastless track plate
CN109883654A (en) * 2019-01-25 2019-06-14 武汉精立电子技术有限公司 A kind of chessboard trrellis diagram, generation method and localization method for OLED sub-pixel positioning
CN110009682A (en) * 2019-03-29 2019-07-12 北京理工大学 A kind of object recognition and detection method based on monocular vision
CN109901594A (en) * 2019-04-11 2019-06-18 清华大学深圳研究生院 A kind of localization method and system of weed-eradicating robot
CN110297498A (en) * 2019-06-13 2019-10-01 暨南大学 A kind of rail polling method and system based on wireless charging unmanned plane
CN110490826A (en) * 2019-08-26 2019-11-22 河南农业大学 Droplet sample image processing method in a kind of camera
CN110490826B (en) * 2019-08-26 2021-09-17 河南农业大学 Fog drop sample image processing method in camera
CN110517325A (en) * 2019-08-29 2019-11-29 的卢技术有限公司 The vehicle body surrounding objects localization method and system of a kind of coordinate transform and coordinate transform
CN110853080A (en) * 2019-09-30 2020-02-28 广西慧云信息技术有限公司 Method for measuring size of field fruit
CN110737942A (en) * 2019-10-12 2020-01-31 清华四川能源互联网研究院 Underwater building model establishing method, device, equipment and storage medium
CN110737942B (en) * 2019-10-12 2023-05-02 清华四川能源互联网研究院 Underwater building model building method, device, equipment and storage medium
CN110796709A (en) * 2019-10-29 2020-02-14 上海眼控科技股份有限公司 Method and device for acquiring size of frame number, computer equipment and storage medium
CN110816522A (en) * 2019-11-12 2020-02-21 深圳创维数字技术有限公司 Vehicle attitude control method, apparatus, and computer-readable storage medium
CN111192235B (en) * 2019-12-05 2023-05-26 中国地质大学(武汉) Image measurement method based on monocular vision model and perspective transformation
CN111192235A (en) * 2019-12-05 2020-05-22 中国地质大学(武汉) Image measuring method based on monocular vision model and perspective transformation
CN111105488B (en) * 2019-12-20 2023-09-08 成都纵横自动化技术股份有限公司 Imaging simulation method, imaging simulation device, electronic equipment and storage medium
CN111105488A (en) * 2019-12-20 2020-05-05 成都纵横自动化技术股份有限公司 Imaging simulation method and device, electronic equipment and storage medium
CN111191646B (en) * 2019-12-23 2023-04-18 福建亿榕信息技术有限公司 Intelligent identification method for pointer instrument
CN111191646A (en) * 2019-12-23 2020-05-22 福建亿榕信息技术有限公司 Intelligent identification method for pointer instrument
CN111568222A (en) * 2020-02-29 2020-08-25 佛山市云米电器科技有限公司 Water dispenser control method, water dispenser and computer readable storage medium
CN111623776A (en) * 2020-06-08 2020-09-04 昆山星际舟智能科技有限公司 Method for measuring distance of target by using near infrared vision sensor and gyroscope
CN112197708A (en) * 2020-08-31 2021-01-08 深圳市慧鲤科技有限公司 Measuring method and device, electronic device and storage medium
CN112258455A (en) * 2020-09-28 2021-01-22 上海工程技术大学 Detection method for detecting spatial position of part based on monocular vision
CN112614191B (en) * 2020-12-16 2024-05-24 江苏智库智能科技有限公司 Loading and unloading position detection method, device and system based on binocular depth camera
CN112614191A (en) * 2020-12-16 2021-04-06 江苏智库智能科技有限公司 Loading and unloading position detection method, device and system based on binocular depth camera
CN112798812A (en) * 2020-12-30 2021-05-14 中山联合汽车技术有限公司 Target speed measuring method based on monocular vision
CN112798812B (en) * 2020-12-30 2023-09-26 中山联合汽车技术有限公司 Target speed measuring method based on monocular vision
CN113222891B (en) * 2021-04-01 2023-12-22 东方电气集团东方锅炉股份有限公司 Line laser-based binocular vision three-dimensional measurement method for rotating object
CN113222891A (en) * 2021-04-01 2021-08-06 东方电气集团东方锅炉股份有限公司 Line laser-based binocular vision three-dimensional measurement method for rotating object
CN113344998B (en) * 2021-06-25 2022-04-29 北京市商汤科技开发有限公司 Depth detection method and device, computer equipment and storage medium
WO2022267275A1 (en) * 2021-06-25 2022-12-29 北京市商汤科技开发有限公司 Depth detection method, apparatus and device, storage medium, computer program and product
CN113344998A (en) * 2021-06-25 2021-09-03 北京市商汤科技开发有限公司 Depth detection method and device, computer equipment and storage medium
CN113592934B (en) * 2021-06-29 2024-02-06 浙江工业大学 Target depth and height measuring method and device based on monocular camera
CN113592934A (en) * 2021-06-29 2021-11-02 浙江工业大学 Monocular camera-based target depth and height measuring method and device
CN113793388A (en) * 2021-08-09 2021-12-14 南京工程学院 Stereoscopic vision interpersonal safe distance detection method based on deep learning
CN113496528A (en) * 2021-09-07 2021-10-12 湖南众天云科技有限公司 Method and device for calibrating position of visual detection target in fixed traffic roadside scene
CN113538546A (en) * 2021-09-17 2021-10-22 智道网联科技(北京)有限公司 Target detection method, device and equipment for automatic driving
CN114022378A (en) * 2021-11-04 2022-02-08 中天智能装备有限公司 Copper strip shielding layer overlapping rate detection method based on vision
CN114155290A (en) * 2021-11-18 2022-03-08 合肥富煌君达高科信息技术有限公司 System and method for large-field-of-view high-speed motion measurement
CN114155290B (en) * 2021-11-18 2022-09-09 合肥富煌君达高科信息技术有限公司 System and method for large-field-of-view high-speed motion measurement
CN114897784B (en) * 2022-04-13 2023-02-21 广东工业大学 Monocular egg size assembly line measuring method
CN114897784A (en) * 2022-04-13 2022-08-12 广东工业大学 Monocular egg size assembly line measuring method
CN114549666A (en) * 2022-04-26 2022-05-27 杭州蓝芯科技有限公司 AGV-based panoramic image splicing calibration method
CN114549666B (en) * 2022-04-26 2022-09-06 杭州蓝芯科技有限公司 AGV-based panoramic image splicing calibration method
CN115026828A (en) * 2022-06-23 2022-09-09 池州市安安新材科技有限公司 Robot arm grabbing control method and system
CN115272110A (en) * 2022-07-21 2022-11-01 四川大学 Projector distortion correction method and device in structured light three-dimensional reconstruction
CN116402871A (en) * 2023-03-28 2023-07-07 苏州大学 Monocular distance measurement method and system based on scene parallel elements and electronic equipment
CN116402871B (en) * 2023-03-28 2024-05-10 苏州大学 Monocular distance measurement method and system based on scene parallel elements and electronic equipment
CN116091608A (en) * 2023-04-11 2023-05-09 深之蓝海洋科技股份有限公司 Positioning method and positioning device for underwater target, underwater equipment and storage medium
CN116758171A (en) * 2023-08-21 2023-09-15 武汉中导光电设备有限公司 Imaging system pose correction method, device, equipment and readable storage medium
CN116758171B (en) * 2023-08-21 2023-10-27 武汉中导光电设备有限公司 Imaging system pose correction method, device, equipment and readable storage medium

Also Published As

Publication number Publication date
CN109035320B (en) 2021-08-10

Similar Documents

Publication Publication Date Title
CN109035320A (en) Depth extraction method based on monocular vision
CN109146980A (en) The depth extraction and passive ranging method of optimization based on monocular vision
CN109269430B (en) Multi-standing-tree breast height diameter passive measurement method based on deep extraction model
Chen et al. High-accuracy multi-camera reconstruction enhanced by adaptive point cloud correction algorithm
CN102376089B (en) Target correction method and system
CN109255818A (en) A kind of extracting method of novel target and its sub-pixel angle point
CN103499297B (en) A kind of high-precision measuring method based on CCD
CN109859272B (en) Automatic focusing binocular camera calibration method and device
CN105716542B (en) A kind of three-dimensional data joining method based on flexible characteristic point
CN109579695B (en) Part measuring method based on heterogeneous stereoscopic vision
CN107977996B (en) Space target positioning method based on target calibration positioning model
CN112132908B (en) Camera external parameter calibration method and device based on intelligent detection technology
CN110672020A (en) Stand tree height measuring method based on monocular vision
CN104463791A (en) Fisheye image correction method based on spherical model
CN103473771A (en) Method for calibrating camera
CN107084680A (en) A kind of target depth measuring method based on machine monocular vision
CN110889829A (en) Monocular distance measurement method based on fisheye lens
CN110807815B (en) Quick underwater calibration method based on corresponding vanishing points of two groups of mutually orthogonal parallel lines
CN109974618B (en) Global calibration method of multi-sensor vision measurement system
CN109448043A (en) Standing tree height extracting method under plane restriction
CN113119129A (en) Monocular distance measurement positioning method based on standard ball
Wang et al. Error analysis and improved calibration algorithm for LED chip localization system based on visual feedback
CN107680035B (en) Parameter calibration method and device, server and readable storage medium
Mei et al. Monocular vision for pose estimation in space based on cone projection
CN102881040A (en) Three-dimensional reconstruction method for mobile photographing of digital camera

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant