CN108765444A - Ground T shape Moving objects detection and location methods based on monocular vision - Google Patents

Ground T shape Moving objects detection and location methods based on monocular vision Download PDF

Info

Publication number
CN108765444A
CN108765444A CN201810520560.8A CN201810520560A CN108765444A CN 108765444 A CN108765444 A CN 108765444A CN 201810520560 A CN201810520560 A CN 201810520560A CN 108765444 A CN108765444 A CN 108765444A
Authority
CN
China
Prior art keywords
target
picture
camera
region
coordinate system
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201810520560.8A
Other languages
Chinese (zh)
Inventor
侯谊
贺风华
姚郁
马杰
郝宁
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Harbin Institute of Technology
Original Assignee
Harbin Institute of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harbin Institute of Technology filed Critical Harbin Institute of Technology
Priority to CN201810520560.8A priority Critical patent/CN108765444A/en
Publication of CN108765444A publication Critical patent/CN108765444A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/136Segmentation; Edge detection involving thresholding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

Ground T shape Moving objects detection and location methods based on monocular vision, are related to target detection technique field.The present invention can accurately obtain the position and direction of T-type plate on ground in order to realize on rotor wing unmanned aerial vehicle.The present invention acquires a certain number of target positive samples and negative sample picture first, and grader is generated according to picture;Then, the picture of input camera acquisition obtains the different region of picture with a certain size detection window and feature is inputted trained grader to the extracted region HOG features of acquisition, to judge whether the region is target;To the region of then target, color notation conversion space is carried out, image segmentation is carried out according to threshold value, is generated and binary map, shape analysis is carried out to binary map later, to obtain the direction of T-type plate;Finally, it according to the posture of camera and elevation information, is coordinately transformed, obtains relative position and the direction of target and aircraft.The present invention is suitable for the relative position and the direction of motion that unmanned plane obtains mobile surface targets.

Description

Ground T shape Moving objects detection and location methods based on monocular vision
Technical field
The method that the present invention relates to a kind of to detect on rotor wing unmanned aerial vehicle and position ground target, is related to target detection Technical field.
Background technology
Rotor unmanned aircraft is a kind of novel in appearance, and the vertical take-off and landing unmanned aerial vehicle of superior performance has structure letter The features such as single, operation is flexibly, load capacity is strong has in military, civil fields such as reconnaissance patrol, the monitoring of natural calamity and supports Have a wide range of applications.
For unmanned plane, the detection of ground target is to realize the basis of unmanned plane and environmental interaction with positioning, can be with Promote automation and the intelligent level of flight control.Task phase is executed in unmanned plane, and detection and positioning to target can be with Improve the ability that unmanned plane independently executes task.
For target detection problems, there is the technology of some.There are two main classes for the method for target detection, first, based on fortune The detection mode of dynamic information, such as background subtraction, but the application scenarios of this kind of method are static background.And the camera shooting of aircraft Head constantly moves, therefore background also changes in real time, and target is also moving.Therefore, mesh is detected using this method There are larger difficulty for mark.Second method is to utilize the feature of target inherently, and target is extracted from background.Such as profit With color characteristic, edge feature is split target.Color, edge, texture are all the intrinsic features in objective world, can The difference between target and background is depicted well.Coloured image contains abundant colouring information, and is easy to apply, and has good Good real-time, but the variation of many factors such as characteristic of illumination, shade, video camera itself, all can be to the segmentation of color characteristic It impacts.Edge is the high-level feature obtained to the integrated treatment of more pixels in image, can describe object well The profile information of itself, and the algorithm of target detection based on edge has good accuracy and robustness, but calculation amount is larger.
Invention content
The present invention proposes a kind of ground T shape Moving objects detection and location methods based on monocular vision, to realize The position and direction of T-type plate on ground can be accurately obtained on rotor wing unmanned aerial vehicle, improve T shape Moving objects detection and locations Accuracy rate.T-type plate of the present invention is as shown in Figure 2.
The present invention adopts the technical scheme that solve above-mentioned technical problem:
Ground T shape Moving objects detection and location methods based on monocular vision, include the following steps:
Step 1: acquiring a certain number of positive samples and negative sample, to the sample extraction HOG features of acquisition, and mark is assigned Feature is sent into label in SVM (support vector machines) and is trained by label, preserves the grader that training obtains;
Step 2: the picture that camera acquires on input unmanned plane, the camera being connected with unmanned plane is just lower to be regarded to obtain Camera height off the ground acquires distance of the camera away from target according to known target height off the ground;
According to the size l in known target real spacex*lyWith the f of height h and fixed-focus away from camera of video camerax, fy, the size of target in the picture is found out, pixel number of the target in image x-axis, y-axis is:
The region of the window interception picture of predefined size is used later;
fxIndicate the focal length in x-axis of the fixed-focus away from camera, unit is pixel;fyIndicate the coke in y-axis of the fixed-focus away from camera Away from unit is pixel;
Step 3: the extracted region HOG features to intercepting picture in step 2, the classification that feature input step one is generated In device, grader exports whether the region is target;
Step 4: to the region of then target, carries out color notation conversion space and carried out based on threshold value according to colouring information Image segmentation generates binary map, by T-type plate and background separation;
Step 5: according to the segmentation result in binary map, rectangle fitting is carried out, T-type plate is determined with rectangle, according to The length side information of fitted rectangle obtains the accurate location of target and possible direction, and according to the face at four angles in rectangular area Color information judges the accurate direction of T-type plate;
Step 6: the position according to the target of step 3 determination in the picture, solves position of the target relative to aircraft; Camera is fixed on aircraft, so the posture information and elevation information of camera can be obtained according to the posture and height of aircraft , according to camera forming model and rotation transformation relationship, the relative position relation of target and aircraft can be acquired.
Further, step 1 includes the following steps:
Step 1 A, a certain number of positive samples of acquisition and negative sample, to the sample extraction HOG features of acquisition, and assign mark Label;
The Target Photo under different angle difference light conditions is acquired, is intercepted comprising complete object just using sectional drawing tool Square region, wherein target percentage in figure is more than 60%, assigns these pictures to label value 1, is for positive sample; Picture of the acquisition not comprising target assigns these pictures to label value -1, is for negative sample;Positive sample and the ratio of negative sample are 1:3;All pictures are adjusted to onesize;
Step 1 B, HOG features are extracted to picture, process is as follows:Color is carried out to input picture using Gamma correction methods The standardization in space;The Grad dx and dy in each pixel level direction and vertical direction in image are calculated later, and calculate ladder The direction θ of degree:
Dx (x, y)=I (x+1, y)-I (x-1, y)
Dy (x, y)=I (x, y+1)-I (c, y-1)
In formula, I (x, y) is that coordinate is the gray value of the pixel of (x, y) in image;
Picture is divided into n*n cell, each region includes m*m pixel;The Gradient Features histogram of generation unit lattice Figure, is divided into k section by gradient direction, and using these sections as the horizontal axis of histogram, the value in each section is that this belongs to the area Between the sum of pixel gradient value, the value in each section in statistic unit lattice is the longitudinal axis of histogram, then each cell includes k A characteristic value;The feature of j*j cell is connected in series together, j<N generates a block B, returns to the feature vector in block One changes, and generates j*j*k feature vector;According to the block feature of default step-length motion scan picture, step-length is one or more single First lattice, the feature of all pieces of statistics, as final HOG features;
Step 1 C, training grader:The picture feature vector that the picture tag that step 1 A is generated is generated with step 1 B It is input in SVM classifier and is trained, generate the grader for classification.
Further, in step 2:
By the length of side size setting w=min (n of target in the picturex,ny), the length of side size of estimation target in the picture is Between w/2 to 2*w;
The flow for intercepting picture region is as follows:First, the w/2 that the size of detection window is set to the target length of side, according to one Fixed step-length moves detection window over an input image, and under normal circumstances, step-length may be set to the 1/2 or 1/ of the detection window length of side 4, the region that window each time is truncated to is preserved, until detection window all carried out interception to input picture full figure;So Afterwards, detection window is amplified to certain multiple, multiple may be set to 1.1 times, continue to move according to certain step-length, until time Go through complete graph;Continuous amplification detection window, intercepts picture region, until the detection window length of side is more than the target maximum side of estimation Long, as w*2 stops interception picture, the region of all interceptions is preserved.
Further, step 4 includes the following steps:
According to the target area that step 3 obtains, corresponding region in original picture is intercepted, color is carried out to the picture of interception Spatial alternation;There is RGB color to be converted to hsv color space in picture according to following formula:
V=max (R, G, B)
if(H<0) H=H+360.
Wherein, R, G, B are pixel in r, g, the numerical value on the channels b, H, S, and V is pixel in h, s, the numerical value on the channels v.
For red and blue target, using different Color Channels:
Red T-type plate is split using the channels H, for blue target, is split using the channels V;For interception Pixel in picture, pixel value meet set interval (Il,Ih) be target point, other point be background dot;By target The pixel value of point is set as 255, the pixel value of background dot is set as 0, you can the binary map after being divided:
Further, in step 5,
According to the binary map for only including T-shaped plate of generation, the minimum enclosed rectangle of T-type plate is obtained using rectangle fitting;T-type The length-width ratio of plate is 5:3, the direction of motion and short side of robot are assured that by the information of boundary rectangle long side and short side Straight line where AC is consistent;Two round a and circle b and T-shaped plate on the upside of T-type plate are not intersections, and two following circles Circle c and circle d are completely superposed with T-shaped plate;The number of the stain and white point in binary map in four circles of T-shaped plate is counted come the side of determination To directions question:
Number(a)+Number(b)>Number (c)+Number (d),
Number (x) indicates the white point number for including in circle x in binary map, thereby determines that the direction of motion has C to be directed toward A.
Further, the process of acquisition aiming spot information is:
Coordinate of the known target point in pixel coordinate system is (u, v);The difference in height h, h that target point is sat with camera>0, it can It is acquired by ultrasonic ranging;Transformational relation between each coordinate system of camera coordinates system, body coordinate system, world coordinate system can be by Spin matrix indicates;
According to camera imaging model, coordinate of the target point under camera coordinates system is acquired as the following formula:
Wherein, s is a dimension scale, and M is Intrinsic Matrix, [Xca Yca Zca]TIt is target point under camera coordinates system Coordinate;
Image coordinate system and body coordinate system, the transformational relation between body coordinate system and competition field coordinate system can be by following public affairs Formula acquires:
Wherein,For the spin matrix of image coordinate system and body coordinate system, R is body coordinate system and competition field coordinate system Between spin matrix;Then by the above transformational relation, can acquire:
Competition field moving coordinate system Z axis straight down, it follows that Z=h;According to transformational relation required before, can obtain Following formula:
M,R it is known that can find out the value of s, and then acquire X, Y as a result, as aiming spot information.
The invention has the advantages that:
The present invention is a kind of method detected using monocular cam and position red or green T-type plate, in rotor It is detected on unmanned plane and positions mobile surface targets.The present invention acquires a certain number of target positive samples and negative sample figure first Piece generates grader according to picture;Then, the picture of input camera acquisition obtains picture with a certain size detection window Feature is inputted trained grader by different regions to the extracted region HOG features of acquisition, whether to judge the region For target;To the region of then target, color notation conversion space is carried out, image segmentation is carried out according to threshold value, is generated and binary map, it Shape analysis is carried out to binary map afterwards, to obtain the direction of T-type plate;Finally, it according to the posture of camera and elevation information, is sat Mark transformation, obtains relative position and the direction of target and aircraft.The present invention is suitable for unmanned plane and obtains mobile surface targets Relative position and the direction of motion.The present invention has higher accuracy rate, is embodied in:
1,90% or more accuracy rate can be reached for the detection of T-type plate on ground
2, the present invention is less than 10 ° to the evaluated error in T-type plate direction, and site error is less than 3cm
3, it is intel i7-6500u using processor, the computer of dominant frequency 2.6GHz is run, the method for the present invention place per second It is about 5 frames to manage speed, meets requirement of real-time.
4, in conjunction with the elevation information of aircraft, detection algorithm algorithm is optimized, improves the efficiency of detection.
Description of the drawings
Fig. 1 is the flow chart of embodiment;
Fig. 2 is T-type plate to be detected;
Fig. 3 is that schematic diagram is extracted in T-type plate direction;
Fig. 4 is the green positive sample picture of acquisition;
Fig. 5 is the red positive sample picture of acquisition;
Fig. 6 is the shift process for resolving position.
Specific implementation mode
The step of a kind of ground T shape Moving objects detection and location methods based on monocular vision of described in present embodiment It is as follows:
Step 1: acquiring a certain number of positive samples and negative sample, to the sample extraction HOG features of acquisition, and mark is assigned Feature is sent into label in SVM (support vector machines) and is trained by label, preserves the grader that training obtains;
Step 2: the picture that camera acquires on input unmanned plane, camera just regard down, and are connected with unmanned plane, so We can obtain camera height off the ground, simultaneously, it is assumed that target height off the ground, can be in the hope of camera away from target it is known that therefore Distance.And UAV Attitude variation range is smaller, therefore, it is just to regard down that can be approximately considered camera.In target real space Size lx*ly, it was known that with the height h of video camera it is known that and we using fixed-focus away from camera, so fx,fy, it was known that by This can find out the size of target in the picture, and pixel number of the target in image x-axis, y-axis is:
The region of the window interception picture of particular size is used later;
Step 3: to the extracted region HOG features of step 4 interception, in the grader that feature input step one is generated, Grader exports whether the region is target;
Step 4: to the region of then target, carries out color notation conversion space and carried out based on threshold value according to colouring information Image segmentation generates binary map, by T-type plate and background separation;
Step 5: according to the segmentation result in binary map, rectangle fitting is carried out, T-type plate is determined with rectangle, according to The length side information of fitted rectangle obtains the accurate location of target and possible direction, and according to the face at four angles in rectangular area Color information judges the accurate direction of T-type plate;
Step 6: the coordinate according to the target of step 3 determination in the picture, solves position of the target relative to aircraft. Camera is fixed on aircraft, so the posture information and elevation information of camera can be obtained according to the posture and height of aircraft , according to camera forming model and rotation transformation relationship, the relative position relation of target and aircraft can be acquired.
Each step is described further:
Step 1 A, a certain number of positive samples of acquisition and negative sample, to the sample extraction HOG features of acquisition, and assign mark Label.First, the Target Photo under different angle difference light conditions is acquired, using sectional drawing tool, interception includes complete object Square area, wherein target percentage in figure is more than 60%, and such as Fig. 4, these pictures are assigned to label shown in Fig. 5 Value 1 is for positive sample;Picture of the acquisition not comprising target assigns these pictures to label value -1, is for negative sample.Positive sample Ratio with negative sample is about 1:3.Can be according to by this onesize step that is adjusted to of all pictures, later by image carry out by According to the processing of following formula gray processing.
Step 1 B, HOG features are extracted to picture.Input picture can be carried out using Gamma correction methods according to the following steps The standardization of color space;The Grad dx and dy in each pixel level direction and vertical direction in image are calculated later, and are counted Calculate the direction θ of gradient:
Dx (x, y)=I (x+1, y)-I (x-1, y)
Dy (x, y)=I (x, y+1)-I (x, y-1)
In formula, I (x, y) is that coordinate is the gray value of the pixel of (x, y) in image.
Picture is divided into n*n cell, each region includes m*m pixel;The Gradient Features histogram of generation unit lattice Figure, is divided into k section by gradient direction, and using these sections as the horizontal axis of histogram, the value in each section is that this belongs to the area Between the sum of pixel gradient value, the value in each section in statistic unit lattice is the longitudinal axis of histogram, then each cell includes k A characteristic value;The feature of j*j cell is connected in series together, j<N generates a block B, returns to the feature vector in block One changes, and generates j*j*k feature vector;According to the block feature of certain step-length motion scan picture, step-length is several cells, Count all pieces of feature, as final HOG features.
Step 1 C, training grader.The picture feature vector that the picture tag that step 1 A is generated is generated with step 1 B It is input in SVM classifier and is trained, generate the grader for classification.
Step 2: the picture that camera acquires on input unmanned plane, camera just regard down, and are connected with unmanned plane, so We can obtain camera height off the ground, simultaneously, it is assumed that target height off the ground, can be in the hope of camera away from target it is known that therefore Distance.And UAV Attitude variation range is smaller, therefore, it is just to regard down that can be approximately considered camera.In target real space Size lx*ly, it was known that with the height h of video camera it is known that and we using fixed-focus away from camera, so fx,fy, it was known that by This can find out the size of target in the picture, and pixel number of the target in image x-axis, y-axis is:
By the length of side size setting w=min (n of target in the picturex,ny).Therefore, we estimate target in the picture Length of side size is between w/2 to 2*w.
The flow for intercepting picture region is as follows.First, the w/2 that the size of detection window is set to the target length of side, according to one Fixed step-length moves detection window over an input image, and under normal circumstances, step-length may be set to the 1/2 or 1/ of the detection window length of side 4, the region that window each time is truncated to is preserved, until detection window all carried out interception to input picture full figure;So Afterwards, detection window is amplified to certain multiple, multiple can be set as 1.1 times, continue to move according to certain step-length, until Traverse complete graph;Continuous amplification detection window, intercepts picture region, until the detection window length of side is more than the target maximum side of estimation Long, as w*2 stops interception picture, the region of all interceptions is preserved.
Step 3: according to the HOG feature extracting methods of step 1 B, HOG features are carried out to the region that step 2 generates and are carried It takes, generates the feature vector in the region, feature vector is input in the grader that step 1 C is generated, grader exports target Class label, if label be 1, the region be target area, if output label be -1, the region be background.
Step 4: according to the target area that step 3 obtains, intercept corresponding region in original picture, to the picture of interception into Row color notation conversion space.Picture is had RGB color to be converted to hsv color space by we, can be according to following formula:
V=max (R, G, B)
if(H<0) H=H+360.
For red and blue target, using different Color Channels.Red T-type plate is split using the channels H, right In blue target, it is split using the channels V.For the pixel in interception picture, pixel value meets set interval (Il, Ih) be target point, other point be background dot.The pixel value of target point is set as 255, the pixel value of background dot is set It is set to 0, you can the binary map after being divided.
Step 5: according to the binary map for only including T-shaped plate generated in step 4, we obtain T-type using rectangle fitting The minimum enclosed rectangle of plate.The length-width ratio of T-type plate is 5:3, therefore can be true by boundary rectangle long side and the information of short side The direction of motion for determining robot is consistent with the straight line where short side AC, as shown in Figure 3.The two round a and circle b and T shapes of upside Plate is not intersection, and following two round c and circle d are completely superposed with T-shaped plate.We can count in binary map The number of stain and white point in four circles of T-shaped plate, so that it may to determine the directions question in direction.
Number(a)+Number(b)>Number(c)+Number(d)
(Number (x) indicates the white point number for including in circle x in binary map) thereby determines that the direction of motion has C to be directed toward A.
Step 6: coordinate of the known target point in pixel coordinate system is (u, v);The difference in height h that target point is sat with camera (h>0) it, can be acquired by ultrasonic ranging;Camera coordinates system, body coordinate system turn between each coordinate system such as world coordinate system The relationship of changing can be indicated that transfer process is as shown in Figure 6 by spin matrix.
According to camera imaging model, coordinate of the target point under camera coordinates system can be acquired by formula.
Wherein, s is a dimension scale, and M is Intrinsic Matrix, [Xca Yca Zca]TIt is target point under camera coordinates system Coordinate.
Image coordinate system and body coordinate system, the transformational relation between body coordinate system and competition field coordinate system can be by following public affairs Formula acquires:
Wherein,For the spin matrix of image coordinate system and body coordinate system, R is body coordinate system and competition field coordinate system Between spin matrix.Then by the above transformational relation, can acquire:
Competition field moving coordinate system Z axis straight down, it follows that Z=h.According to transformational relation required before, can obtain Following formula:
M,R it is known that can find out the value of s, and then acquire X, Y as a result, as aiming spot information.

Claims (6)

1. the ground T shape Moving objects detection and location methods based on monocular vision, which is characterized in that include the following steps:
Step 1: acquiring a certain number of positive samples and negative sample, to the sample extraction HOG features of acquisition, and label is assigned, it will Feature is sent into SVM with label and is trained, and the grader that training obtains is preserved;
Step 2: the picture that camera acquires on input unmanned plane, the camera being connected with unmanned plane is just lower to be regarded to be imaged Head height off the ground, distance of the camera away from target is acquired according to known target height off the ground;
According to the size l in known target real spacex*lyWith the f of height h and fixed-focus away from camera of video camerax,fy, ask Go out the size of target in the picture, pixel number of the target in image x-axis, y-axis is:
The region of the window interception picture of predefined size is used later;
fxIndicate the focal length in x-axis of the fixed-focus away from camera, unit is pixel;fyIndicate the focal length in y-axis of the fixed-focus away from camera, it is single Position is pixel;
Step 3: the extracted region HOG features to intercepting picture in step 2, the grader that feature input step one is generated In, grader exports whether the region is target;
Step 4: to the region of then target, carries out color notation conversion space and the image based on threshold value is carried out according to colouring information Segmentation generates binary map, by T-type plate and background separation;
Step 5: according to the segmentation result in binary map, rectangle fitting is carried out, T-type plate is determined with rectangle, according to fitting The length side information of rectangle obtains the accurate location of target and possible direction, and is believed according to the color at the angle of rectangular area four Breath, judges the accurate direction of T-type plate;
Step 6: the position according to the target of step 3 determination in the picture, solves position of the target relative to aircraft;It will take the photograph As head is fixed on aircraft, so the posture information and elevation information of camera can be obtained according to the posture and height of aircraft, According to camera forming model and rotation transformation relationship, the relative position relation of target and aircraft can be acquired.
2. the ground T shape Moving objects detection and location methods according to claim 1 based on monocular vision, feature exist In step 1 includes the following steps:
Step 1 A, a certain number of positive samples of acquisition and negative sample, to the sample extraction HOG features of acquisition, and assign label;
The Target Photo under different angle difference light conditions is acquired, the square of complete object is included using the interception of sectional drawing tool Region, wherein target percentage in figure is more than 60%, assigns these pictures to label value 1, is for positive sample;Acquisition Picture not comprising target assigns these pictures to label value -1, is for negative sample;Positive sample and the ratio of negative sample are 1:3; All pictures are adjusted to onesize;
Step 1 B, HOG features are extracted to picture, process is as follows:Color space is carried out to input picture using Gamma correction methods Standardization;The Grad dx and dy in each pixel level direction and vertical direction in image are calculated later, and calculate gradient Direction θ:
Dx (x, y)=I (x+1, y)-I (x-1, y)
Dy (x, y)=I (x, y+1)-I (x, y-1)
In formula, I (x, y) is that coordinate is the gray value of the pixel of (x, y) in image;
Picture is divided into n*n cell, each region includes m*m pixel;The Gradient Features histogram of generation unit lattice, will Gradient direction is divided into k section, and using these sections as the horizontal axis of histogram, the value in each section is the picture for belonging to the section The sum of plain Grad, the value in each section in statistic unit lattice, is the longitudinal axis of histogram, then each cell includes k feature Value;The feature of j*j cell is connected in series together, j<N generates a block B, the feature vector in block is normalized, Generate j*j*k feature vector;According to the block feature of default step-length motion scan picture, step-length is one or more cells, Count all pieces of feature, as final HOG features;
Step 1 C, training grader:The picture tag that step 1 A is generated is inputted with the picture feature vector that step 1 B is generated To being trained in SVM classifier, the grader for classification is generated.
3. the ground T shape Moving objects detection and location methods according to claim 1 based on monocular vision, feature exist In in step 2:
By the length of side size setting w=min (n of target in the picturex,ny), the length of side size of estimation target in the picture is w/2 To between 2*w;
The flow for intercepting picture region is as follows:First, the w/2 that the size of detection window is set to the target length of side, according to certain Step-length moves detection window over an input image, and under normal circumstances, step-length may be set to the 1/2 or 1/4 of the detection window length of side, will The region that window is truncated to each time preserves, until detection window all carried out interception to input picture full figure;Then, will Detection window is amplified to certain multiple, and multiple may be set to 1.1 times, continues to move according to certain step-length, until having traversed Full figure;Continuous amplification detection window, intercepts picture region, until target maximal side of the detection window length of side more than estimation, i.e., For w*2, stops interception picture, the region of all interceptions is preserved.
4. the ground T shape Moving objects detection and location methods according to claim 1 based on monocular vision, feature exist In step 4 includes the following steps:
According to the target area that step 3 obtains, corresponding region in original picture is intercepted, color space is carried out to the picture of interception Transformation;There is RGB color to be converted to hsv color space in picture according to following formula:
V=max (R, G, B)
if(H<0) H=H+360.
Wherein, R, G, B are pixel in r, g, the numerical value on the channels b, H, S, and V is pixel in h, s, the numerical value on the channels v.
For red and blue target, using different Color Channels:
Red T-type plate is split using the channels H, for blue target, is split using the channels V;For intercepting picture In pixel, pixel value meets set interval (Il,Ih) be target point, other point be background dot;By target point Pixel value is set as 255, the pixel value of background dot is set as 0, you can the binary map after being divided:
5. the ground T shape Moving objects detection and location methods according to claim 1 based on monocular vision, feature exist In, in step 5,
According to the binary map for only including T-shaped plate of generation, the minimum enclosed rectangle of T-type plate is obtained using rectangle fitting;T-type plate Length-width ratio is 5:3, the direction of motion and the short side AC institutes of robot are assured that by the information of boundary rectangle long side and short side Straight line it is consistent;Two round a on the upside of T-type plate and circle b and T-shaped plate are not intersections, and following two round c and Circle d is completely superposed with T-shaped plate;The number of the stain and white point in binary map in four circles of T-shaped plate is counted to determine direction Directions question:
Number(a)+Number(b)>Number (c)+Number (d),
Number (x) indicates the white point number for including in circle x in binary map, thereby determines that the direction of motion has C to be directed toward A.
6. the ground T shape Moving objects detection and location methods according to claim 1 based on monocular vision, feature exist In the process for obtaining aiming spot information is:
Coordinate of the known target point in pixel coordinate system is (u, v);The difference in height h, h that target point is sat with camera>0, it can be by surpassing Sound ranging acquires;Transformational relation between each coordinate system of camera coordinates system, body coordinate system, world coordinate system can be by rotating Matrix indicates;
According to camera imaging model, coordinate of the target point under camera coordinates system is acquired as the following formula:
Wherein, s is a dimension scale, and M is Intrinsic Matrix, [Xca Yca Zca]TFor seat of the target point under camera coordinates system Mark;
Image coordinate system and body coordinate system, the transformational relation between body coordinate system and competition field coordinate system can be asked by following formula ?:
Wherein,For the spin matrix of image coordinate system and body coordinate system, R is between body coordinate system and competition field coordinate system Spin matrix;Then by the above transformational relation, can acquire:
Competition field moving coordinate system Z axis straight down, it follows that Z=h;According to transformational relation required before, can obtain following Formula:
M,R it is known that can find out the value of s, and then acquire X, Y as a result, as aiming spot information.
CN201810520560.8A 2018-05-25 2018-05-25 Ground T shape Moving objects detection and location methods based on monocular vision Pending CN108765444A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810520560.8A CN108765444A (en) 2018-05-25 2018-05-25 Ground T shape Moving objects detection and location methods based on monocular vision

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810520560.8A CN108765444A (en) 2018-05-25 2018-05-25 Ground T shape Moving objects detection and location methods based on monocular vision

Publications (1)

Publication Number Publication Date
CN108765444A true CN108765444A (en) 2018-11-06

Family

ID=64005977

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810520560.8A Pending CN108765444A (en) 2018-05-25 2018-05-25 Ground T shape Moving objects detection and location methods based on monocular vision

Country Status (1)

Country Link
CN (1) CN108765444A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109753081A (en) * 2018-12-14 2019-05-14 中国矿业大学 A kind of patrol unmanned machine system in tunnel based on machine vision and air navigation aid
CN111402324A (en) * 2019-01-02 2020-07-10 ***通信有限公司研究院 Target measuring method, electronic equipment and computer storage medium
CN112200866A (en) * 2020-09-30 2021-01-08 无锡太机脑智能科技有限公司 Indoor monocular global positioning method based on color block visual marks

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105021184A (en) * 2015-07-08 2015-11-04 西安电子科技大学 Pose estimation system and method for visual carrier landing navigation on mobile platform
CN105184301A (en) * 2015-09-07 2015-12-23 复旦大学 Method for distinguishing vehicle azimuth by utilizing quadcopter
US20170017240A1 (en) * 2015-07-18 2017-01-19 Tata Consultancy Services Limited Methods and systems for landing of unmanned aerial vehicle

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105021184A (en) * 2015-07-08 2015-11-04 西安电子科技大学 Pose estimation system and method for visual carrier landing navigation on mobile platform
US20170017240A1 (en) * 2015-07-18 2017-01-19 Tata Consultancy Services Limited Methods and systems for landing of unmanned aerial vehicle
CN105184301A (en) * 2015-09-07 2015-12-23 复旦大学 Method for distinguishing vehicle azimuth by utilizing quadcopter

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
YI HOU 等: "《Target detection algorithm for IARC seventh mission》", 《2017 11TH ASIAN CONTROL CONFERENCE》 *
宓超,等: "《装卸机器视觉及其应用》", 31 January 2016 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109753081A (en) * 2018-12-14 2019-05-14 中国矿业大学 A kind of patrol unmanned machine system in tunnel based on machine vision and air navigation aid
CN109753081B (en) * 2018-12-14 2020-08-21 煤炭科学研究总院 Roadway inspection unmanned aerial vehicle system based on machine vision and navigation method
CN111402324A (en) * 2019-01-02 2020-07-10 ***通信有限公司研究院 Target measuring method, electronic equipment and computer storage medium
CN111402324B (en) * 2019-01-02 2023-08-18 ***通信有限公司研究院 Target measurement method, electronic equipment and computer storage medium
CN112200866A (en) * 2020-09-30 2021-01-08 无锡太机脑智能科技有限公司 Indoor monocular global positioning method based on color block visual marks
CN112200866B (en) * 2020-09-30 2024-05-14 无锡太机脑智能科技有限公司 Indoor monocular global positioning method based on color block visual marks

Similar Documents

Publication Publication Date Title
CN110781827B (en) Road edge detection system and method based on laser radar and fan-shaped space division
CN109949361A (en) A kind of rotor wing unmanned aerial vehicle Attitude estimation method based on monocular vision positioning
CN110097553B (en) Semantic mapping system based on instant positioning mapping and three-dimensional semantic segmentation
CN103324936B (en) A kind of vehicle lower boundary detection method based on Multi-sensor Fusion
CN108805906A (en) A kind of moving obstacle detection and localization method based on depth map
CN107392929B (en) Intelligent target detection and size measurement method based on human eye vision model
CN108731587A (en) A kind of the unmanned plane dynamic target tracking and localization method of view-based access control model
CN105447853A (en) Flight device, flight control system and flight control method
CN106780560B (en) Bionic robot fish visual tracking method based on feature fusion particle filtering
CN106874884A (en) Human body recognition methods again based on position segmentation
CN112215860A (en) Unmanned aerial vehicle positioning method based on image processing
CN108765444A (en) Ground T shape Moving objects detection and location methods based on monocular vision
CN108009494A (en) A kind of intersection wireless vehicle tracking based on unmanned plane
CN111126184A (en) Post-earthquake building damage detection method based on unmanned aerial vehicle video
CN114004977A (en) Aerial photography data target positioning method and system based on deep learning
CN107240112A (en) Individual X Angular Point Extracting Methods under a kind of complex scene
CN109101926A (en) Aerial target detection method based on convolutional neural networks
CN111008994A (en) Moving target real-time detection and tracking system and method based on MPSoC
WO2020010620A1 (en) Wave identification method and apparatus, computer-readable storage medium, and unmanned aerial vehicle
CN103955673B (en) Body recognizing method based on head and shoulder model
CN111626241A (en) Face detection method and device
CN104077611B (en) Indoor scene monocular vision space recognition method under class ground gravitational field environment
CN108388854A (en) A kind of localization method based on improvement FAST-SURF algorithms
CN115797397B (en) Method and system for all-weather autonomous following of robot by target personnel
CN113781524B (en) Target tracking system and method based on two-dimensional label

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20181106