CN105139401A - Depth credibility assessment method for depth map - Google Patents
Depth credibility assessment method for depth map Download PDFInfo
- Publication number
- CN105139401A CN105139401A CN201510548522.XA CN201510548522A CN105139401A CN 105139401 A CN105139401 A CN 105139401A CN 201510548522 A CN201510548522 A CN 201510548522A CN 105139401 A CN105139401 A CN 105139401A
- Authority
- CN
- China
- Prior art keywords
- depth
- pixel
- original
- map
- degree
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10028—Range image; Depth image; 3D point clouds
Landscapes
- Image Analysis (AREA)
Abstract
The invention provides a depth credibility assessment method for a depth map. The method comprises the steps: obtaining at least two groups of first original image pairs in different scenes; generating a predictor according to the first original image pairs; obtaining an inputted second original image pair, wherein the second original image pair comprises a second original color map and a second original depth map, which are aligned with each other in one scene; extracting a second characteristic vector of the second original image pair; and predicting the depth credibility of the second original depth map according to the predictor and the second characteristic vector. The method can predict the confidence of scene depth information collected based on low-precision depth equipment.
Description
Technical field
The present invention relates to computer vision field, refer to the appraisal procedure of the confidence level of the degree of depth in depth map especially.
Background technology
At present, at computer vision field, obtain depth information of scene is accurately one of important research content always.Obtain depth information and mainly contain two kinds of methods: the method based on solid geometry theory and the method based on depth transducer.
Obtain depth information of scene based on solid geometry theory is the method that people compare concern all the time, and the method is employed the earliest, and corresponding algorithm is relatively ripe.But its time complexity is higher, what the effect of Depth Information Acquistion depended on scene texture to a great extent enriches degree, thus the considerable restraint range of application of this method.
Along with the development of hardware technology, the scene depth acquisition methods based on hardware sensor have also been obtained increasing concern.This method obtains depth information of scene from single visual angle, and the corresponding feature of obstructed overmatching obtains scene depth, and time complexity is low, and meanwhile, the degree of depth obtains the texture not relying on scene, so there is good robustness.Particularly in recent years, along with selling of low price depth transducer, scene depth is obtained accomplished good unification with price especially, promoted the application of the method greatly.But due to the character of low price depth transducer equipment self, it is than being easier to the impact being subject to noise, and then depth information of scene is caused to catch complete and precision is not high.
In order to the scene depth degree of confidence not high to precision is predicted, and then can provide tutorial message to other application, a lot of method produces.For ToF flying time technology camera, the people such as May and Swadzba calculate the degree of confidence of scene depth according to the amplitude that ToF camera depth transducer equipment obtains.But simple dependence amplitude is foundation is insufficient, and this method robustness is poor, and the degree of confidence of depth information of scene is easily labeled mistake.
The people such as Reynolds propose the degree of confidence of a kind of method prediction ToF camera.They use high accuracy depth sensor device, such as laser 3 d scanner, the depth information of scene gathered is as standard value, by calculating the depth information of scene and the difference of standard value that low side depth transducer obtains, and be translated into degree of confidence, extract the feature of the image that low side depth transducer equipment gathers simultaneously, use the fallout predictor that the method for training obtains, and then predict the degree of confidence of the depth information of scene that low end sensor obtains.This method needs the support of high precision apparatus, thus limits its application to a certain extent; Simultaneously their feature of proposing be not suitable for other depth transducer, such as Kinect (name that Microsoft formally issues body sense periphery peripheral hardware) equipment.
Summary of the invention
The technical problem to be solved in the present invention is, provides the appraisal procedure of the confidence level of the degree of depth in a kind of depth map, can predict the degree of confidence of the depth information of scene gathered based on low precision depth device.
In described depth map, the appraisal procedure of the confidence level of the degree of depth comprises:
Obtain at least two group first original images pair under different scene, described first original image is to the first original color figure and the first original depth-map that comprise alignment;
According to described first original image pair, generation forecast device; Described fallout predictor represents the mapping relations between proper vector and degree of depth confidence level extracted from image;
Obtain the second original image pair of input, described second original image is to the second original color figure alignd under comprising a scene and the second original depth-map;
Extract the second feature vector that described second original image is right;
According to described fallout predictor and described second feature vector, predict the confidence level of the degree of depth of described second original depth-map.
Described according to described first original image pair, the step of generation forecast device comprises:
According at least two group first original images pair described under each scene, calculate average color figure corresponding to scene described in each and average depth map;
According to the described mean depth figure of scene described in each, generate the degree of depth confidence level figure under scene described in each;
According to the described average color figure under scene described in each and described mean depth figure, extract the feature that pixel is associated with degree of depth confidence level, composition first eigenvector
According to described first eigenvector
and the described degree of depth confidence level figure under scene described in each, utilize random forest method, the generation forecast device by training.
Described fallout predictor is:
Wherein, (i, j) represents the coordinate of pixel in described average color figure and described mean depth figure;
refer to the proper vector extracted from image;
F refers to the mapping relations between proper vector and degree of depth confidence level;
C
i, jrefer to the degree of depth confidence level of pixel (i, j).
Described according to the described mean depth figure under scene described in each, the step generating the degree of depth confidence level figure under scene described in each comprises:
According to the described mean depth figure under scene described in each, calculate the variance depth map under scene described in each;
Utilize the Sigmoid tangent bend function after adjustment, the described variance depth map under scene described in each is converted to the degree of depth confidence level figure under scene described in each.
Described according to the described mean depth figure under scene described in each, the step calculating the variance depth map under scene described in each is carried out according to following formula:
Wherein, m is that each organizes the sequence number of the first original depth-map of the first original image centering; The total quantity that the first original image that N gathers under referring to a scene is right, D
ijit is the pixel value of the first original depth-map;
refer to the pixel value of mean depth figure; v
i, jrepresent the variance referring to that pixel (i, j) is corresponding.
Sigmoid tangent bend function after described utilization adjustment, the step of the degree of depth confidence level figure be converted under scene described in each by the described variance depth map under scene described in each is carried out according to following formula:
Wherein, c
i, jrefer to the confidence level of pixel (i, j); v
i, jrefer to the variance of pixel (i, j);
E is natural constant, and λ is conversion parameter.
The step of at least two group first original images under described acquisition scene described in each comprises:
Gather the first original image pair under a scene;
When described first original image is to when comprising N frame the first original depth-map, according to described N frame first original depth-map, calculate the first mean depth figure; Average again after cumulative to the pixel value of all pixels in described N frame first mean depth figure, obtain the first pixel value D;
When described first original image is to when comprising N+1 frame the first original depth-map, according to described N+1 frame first original depth-map, calculate the second mean depth figure; Average again after cumulative to the pixel value of all pixels in described N+1 frame second mean depth figure, obtain the second pixel value D';
Judge that the absolute value whether little 5 of the difference between described first pixel value D and described second pixel value D' is in threshold value;
When the absolute value of the difference between described first pixel value D and described second pixel value D' is less than threshold value, stop gathering the first original image pair under this scenario.
Described first eigenvector comprise following one or more:
The radial distance of image slices vegetarian refreshments X (i, j)
The value of the following feature of image slices vegetarian refreshments X (i, j) and each described feature minimum value in the picture, maximal value, average and intermediate value;
Describedly to be characterized as:
The R red component of image slices vegetarian refreshments X (i, j), G green component, B blue component, D depth component;
The gradient information of pixel (i, j) in mean depth figure;
The marginal information of pixel (i, j) in mean depth figure;
The marginal information of pixel (i, j) in average color figure;
Pixel (i, j) carries out the filtered information in Gabor Garbo to average color figure;
Pixel (i, j) carries out the filtered information in Gabor Garbo to mean depth figure;
The texture information of pixel (i, j) in mean depth figure;
The texture information of pixel (i, j) in average color figure;
Pixel (i, j) arrives the distance of degree of depth hole in mean depth figure.
Described second feature vector comprise following one or more:
The radial distance of image slices vegetarian refreshments X (i, j)
The value of the following feature of image slices vegetarian refreshments X (i, j) and each described feature minimum value in the picture, maximal value, average and intermediate value;
Describedly to be characterized as:
The R red component of image slices vegetarian refreshments X (i, j), G green component, B blue component, D depth component;
The gradient information of pixel (i, j) in the second original depth-map;
The marginal information of pixel (i, j) in the second original depth-map;
The marginal information of pixel (i, j) in the second original color figure;
Pixel (i, j) carries out the filtered information in Gabor Garbo to the second original color figure;
Pixel (i, j) carries out the filtered information in Gabor Garbo to the second original depth-map;
The texture information of pixel (i, j) in the second original depth-map;
The texture information of pixel (i, j) in the second original color figure;
Pixel (i, j) arrives the distance of degree of depth hole in the second original depth-map.
The beneficial effect of technique scheme of the present invention is as follows:
The scene depth figure of the many groups of alignment that the present invention gathers with low accurate facility self and cromogram are for foundation, use random forest method, training fallout predictor, thus the depth image degree of confidence of the effectively low accurate facility acquisition of prediction, and then it can be used as input data to join in other application.
The depth maps of many groups of alignment that disclosure of the invention one utilizes low accurate facility (such as Kinect device) to gather and the method for cromogram predetermined depth figure confidence level.Its basic methods comprises:
First, the cromogram of many group alignment and depth map is used to calculate average color and the depth image of often organizing image;
Then, extract according to average color and depth map the feature be associated with degree of depth confidence level;
Meanwhile, utilize mean depth figure to calculate and often organize the right degree of depth variogram of image, use sigmoid function to be converted into degree of depth confidence level figure;
Then, utilize the feature composition characteristic of said extracted vector and their corresponding degree of depth confidence levels, use random forest method training fallout predictor.
Then, for the colour and the depth image pair that need reliability forecasting, usage forecastings device can the confidence level of effective predetermined depth.Degree of depth confidence level can as input market demand in more senior computer vision, augmented reality personage.
The method of the depth accuracy confidence level that assessment Kinect device of the present invention gathers, the image that this method uses Kinect device to gather is to the method with machine learning, do not need the support of high accuracy depth equipment, such as high-precision laser 3 d scanner, only need the colour after the alignment using Kinect device self to gather and depth image pair, just can judge the confidence level of Kinect depth accuracy, and extract corresponding feature, then machine learning method is utilized to train fallout predictor, when the colour newly gathered and depth image are to input, with regard to the degree of confidence of the degree of depth of its correspondence measurable.
Application scenarios of the present invention is below described.For Kinect device, a kind of method assessing the depth accuracy confidence level that Kinect device gathers, comprises the following steps:
Step 1, chooses the different scenes under indoor environment.Under each scene (visual angle of each scene is different), gather cromogram and the depth map of several alignment by Kinect device, composition multiple image pair.Here alignment parameters can use the parameter stored in Kinect device to carry out.The right quantity of image is as the criterion so that the mean depth figure tried to achieve under this quantity is stable.Mean depth image refers to that each the frame depth image for the multiple series of images centering gathered under a certain scene adds up, and then solves mean value.When image comprises N frame depth image to group, averaging of income depth image is D; When comprising N+1 frame, averaging of income depth image is D', if abs (D-D') <t (abs represents absolute value), (t is a certain threshold value), then think that image log amount is enough stable; Otherwise, continue to gather under this scene.
Step 2, for cromogram and the depth map of several alignment under each scene, calculates their average color figure and average depth map; Then calculate the variance depth map under this scene according to mean depth figure, computing method are as follows:
Wherein, the colour that N gathers under referring to each scene and the right quantity of depth image, D refers to wherein a certain amplitude deepness image,
refer to mean depth image.It is to be noted: gather in image, degree of depth pixel value be zero pixel be not considered at this.
Step 3, variance can reflect the stability of data, and variance little expression data are more stable, otherwise then data are unstable.In [2] step, pixel depth variance is large, represents that depth value obtains unstable, can affect data precision; Pixel depth variance is little, and represent that the degree of depth obtains stable, data precision is higher.Utilize the Sigmoid function after adjustment, degree of depth variogram is converted to the degree of depth confidence level figure of standard, Sigmoid function is as follows:
Wherein, c
i, j, v
i, jrefer to the confidence level (degree of confidence) that pixel (i, j) is corresponding and variance, e is natural constant, and λ can be set to: 0.17;
Step 4, according to the many groups average color figure calculated in step 2 and average depth map, extracts the feature that each pixel of image pair is associated with degree of depth confidence level, composition characteristic vector
feature comprises:
(1) colour that comprises of pixel and depth information, i.e. R (redness), G (green), B (blueness), D (degree of depth) component of image slices vegetarian refreshments X (i, j)
formula is as follows:
Wherein, i, j represent pixel coordinate, I
i, j=(R, G, B), represents the chromatic information of pixel (i, j); D
i, j=D, represents the depth information of pixel (i, j).
(2) radial distance of described pixel (i, j)
refer to the Euclidean distance between current pixel point and image center.Because each pixel has RGBD tetra-kinds of information, RGB is stored in coloured image, and D is stored in depth image, and two kinds of image resolution ratios are consistent, and image center here can refer to any one of two kinds of images.Formula is as follows:
Wherein, (p
i, p
j) refer to the coordinate of pixel (i, j); (c
i, c
j) refer to the coordinate of image center.
(3) gradient information of described pixel (i, j) in depth image, comprises gradient in the x direction
gradient in y-direction
population gradient
according to the intensity magnitude information that Robert Robert operator calculates
and according to the edge direction theta information that Robert Robert operator calculates
respectively according to following formulae discovery:
Wherein,
D (i, j) refers to the depth value that in depth image, pixel (i, j) is corresponding;
D (i+1, j+1) refers to the depth value of pixel (i+1, j+1); D (i+1, j) refers to the depth value of pixel (i+1, j); The depth value of D (i, j+1)
(4) marginal information of described pixel (i, j) in depth image
with the marginal information of described pixel (i, j) in coloured image
the pixel of fringe region is more easily by noise effect, and use the marginal information of Canny card Buddhist nun's operator extraction coloured image and depth image here, formula is as follows:
Wherein, ICanny
i, jand DCnny
i, jthe marginal information of cromogram referring to respectively to use canny algorithm to extract at pixel (i, j) and the marginal information of depth map, use the canny convolution kernel computation bound information of 3 × 3,5 × 5 and 7 × 7 in this method.
(5) pixel (i, j) is at the Gabor Garbo filtering information of depth map
with the Gabor Garbo filtering information in cromogram
In this method, parameter is 0 °, 45 °, 90 °, and the Gabor filter of 135 ° does filtering to coloured image and depth image respectively.Formula is as follows:
Wherein, IGabor
i, jand DGabor
i, jrefer to respectively carry out the filtered result of Gabor to coloured image and depth image;
(6) pixel (i, j) is at the texture information of depth map and cromogram
with
lBP (local binary patterns) algorithm is used to extract texture information to coloured image and depth image respectively in this method.Formula is as follows:
Wherein, ILBP
i, jand DLBP
i, jrefer to that LBP algorithm is applied to the result on coloured image and depth image respectively respectively;
(7) described pixel (i, j) is to the distance of degree of depth hole
degree of depth hole refers to the region that in depth map, depth data cannot gather because of the reason such as light, noise, namely the null value region in depth image, represent that the depth information of scene cannot obtain with depth transducer, so the closer pixel in distance degree of depth cavity is more easily subject to the impact of noise.Formula is as follows:
Wherein, (p
i, p
j) refer to the coordinate of pixel (i, j); (ph
x, ph
y) span is from the coordinate of the nearest degree of depth hole pixel of pixel (i, j).
(8) global characteristics.Because the minimum value in entire image feature, maximal value, average and intermediate value contain important information when stating picture characteristics.So in this method, above-mentioned several features (except the radial distance) minimum value in entire image, maximal value, average and intermediate value are all comprised in proper vector.Wherein, intermediate value is that in one group of data, size is arranged in middle numerical value.
According to above-mentioned feature, each minimum value, maximal value, average and intermediate value comprise 26 dimensional features.Final proper vector comprises 131 dimensional features (26 features × 5+ radial distance).
According to front 8 kinds of features, except radial distance
outward, have 26 dimensional features at each pixel (i, j) place, comprise R, G, B, D,
3 × 3
5 × 5
7 × 7
3 × 3
5 × 5
7 × 7
0 degree
45 degree
90 degree
135 degree
0 degree
45 degree
90 degree
135 degree
on this basis, for the feature that all pixels calculate, the maximal values (26 dimension) can tried to achieve in this dimension all in each dimension, minimum value (26 dimension), average (26 dimension) and intermediate value (26 dimension), so total characteristic dimension totally 26 × 5+1=131 dimension.
About the explanation of profile maxima, minimum value, average and intermediate value, for maximal value, and be only characterized as example with R, G, B, D these four kinds:
Suppose that we have one group of average color figure and average depth map now, comprise four pixels altogether.
The RGBD of pixel 1 correspondence is respectively: 2,4,3,5;
The RGBD of pixel 2 correspondence is respectively: 5,7,8,4;
The RGBD of pixel 3 correspondence is respectively: 4,9,7,9;
The RGBD of pixel 4 correspondence is respectively: 6,3,2,8;
Maximal value then for first dimensional feature (i.e. the R value of each pixel) of this group image is: 6, corresponding two, three, the maximal value of four dimensional features is: 9,8,9, then final proper vector is: R, G, B, D, max (R), max (G), max (B), max (D);
Pixel 1:2,4,3,5,6,9,8,9;
Pixel 2:5,7,8,4,6,9,8,9;
Pixel 3:4,9,7,9,6,9,8,9;
Pixel 4:6,3,2,8,6,9,8,9;
For minimum value, average and intermediate value are adopted and are used the same method.
Step 5, by the degree of depth confidence level calculated in the proper vector extracted in step 4 and step 3, utilizes random forest method, training fallout predictor, and concrete correspondence is as follows:
Wherein,
refer to the proper vector extracted from cromogram and depth map, f refers to the mapping relations between proper vector with corresponding degree of confidence.In the training process, use 100 to set and build random forest, the degree of depth of every tree is 25, and the precision simultaneously set is 0.0001.
That is, for the depth image often organizing alignment and the coloured image of Kinect device collection, corresponding mean depth figure and average cromogram can be calculated, and then the variance depth map of each group data can be obtained.For each pixel value v in variance depth image
i, j, be translated into degree of confidence c by sigmodel function
i, j.This degree of confidence correspond to picture point (i, j).Extract cromogram and feature relevant to degree of depth degree of confidence in depth map, composition characteristic vector, then uses the method based on study simultaneously, training fallout predictor.
Step 6, to the degree of depth and the coloured image pair of new input, extracting the proper vector identical with step 4, training the fallout predictor obtained, predetermined depth confidence level, end-of-job according to entering forest.Because the degree of depth that newly inputs and coloured image are to just one group, therefore, do not need to calculate average color figure and average depth map, directly extract from depth map and cromogram.
The invention provides a kind of based on the degree of depth method for evaluating confidence of low precision (such as: Kinect) equipment self with random forest, the degree of confidence of the depth information of scene that Kinect depth device gathers can be predicted.The scene depth figure of the many groups of alignment that the present invention gathers with Kinect device self and cromogram are for foundation, use random forest method, training fallout predictor, thus the depth image degree of confidence of effectively prediction Kinect device acquisition, and then it can be used as input data to join in other application.That is, by the data gathered, the method training fallout predictor of applied for machines study, the mapping relationship f namely in formula.For the coloured image and the depth image pair that need forecast confidence, extract identical proper vector, usage forecastings device, just can obtain the degree of confidence of image to each pixel.
That is, use above-mentioned feature, composition characteristic vector, simultaneously according to the standard confidence level calculated, use random forest method, train a fallout predictor.Because confidence level is continuous print, thus random forest uses Regression Model.For the colour and the depth image pair that need predetermined depth confidence level, extract identical proper vector, the fallout predictor obtained has been trained in application, obtains their degree of depth confidence level.The present invention does not need the support of high accuracy depth scanning device (as high-precision laser spatial digitizer etc.).
The above is the preferred embodiment of the present invention; it should be pointed out that for those skilled in the art, under the prerequisite not departing from principle of the present invention; can also make some improvements and modifications, these improvements and modifications also should be considered as protection scope of the present invention.
Accompanying drawing explanation
Fig. 1 is the schematic flow sheet of the appraisal procedure of the confidence level of the degree of depth in depth map of the present invention;
Fig. 2 is the principle schematic of the appraisal procedure of the confidence level of the degree of depth in depth map of the present invention.
Embodiment
For making the technical problem to be solved in the present invention, technical scheme and advantage clearly, be described in detail below in conjunction with the accompanying drawings and the specific embodiments.
As shown in Figure 1, in described depth map, the appraisal procedure of the confidence level of the degree of depth comprises:
Step 11, obtains at least two group first original images pair under different scene, and described first original image is to the first original color figure and the first original depth-map that comprise alignment; Scene can be indoor scene.Alignment is under depth image and coloured image are transformed into same visual angle.
Step 12, according to described first original image pair, generation forecast device; Described fallout predictor represents the mapping relations between proper vector and degree of depth confidence level extracted from image;
Step 13, obtains the second original image pair of input, and described second original image is to the second original color figure alignd under comprising a scene and the second original depth-map;
Step 14, extracts the second feature vector that described second original image is right;
Step 15, according to described fallout predictor and described second feature vector, predicts the confidence level of the degree of depth of described second original depth-map.
Wherein, step 12 comprises:
Step 121, according at least two group first original images pair described under each scene, calculates average color figure corresponding to scene described in each and average depth map;
Step 122, according to the described mean depth figure of scene described in each, generates the degree of depth confidence level figure under scene described in each;
Step 123, according to the described average color figure under scene described in each and described mean depth figure, extracts the feature that pixel is associated with degree of depth confidence level, composition first eigenvector
Step 124, according to described first eigenvector
and the described degree of depth confidence level figure under scene described in each, utilize random forest method, the generation forecast device by training.
Described fallout predictor is:
Wherein, (i, j) represents the coordinate of pixel in described average color figure and described mean depth figure;
refer to the proper vector extracted from described average color figure and described mean depth figure;
F refers to the mapping relations between proper vector and degree of depth confidence level;
C
i, jrefer to the degree of depth confidence level of pixel (i, j).
Wherein, step 122 comprises:
Step 1221, according to the described mean depth figure under scene described in each, calculates the variance depth map under scene described in each;
Step 1222, utilizes the Sigmoid tangent bend function after adjustment, the described variance depth map under scene described in each is converted to the degree of depth confidence level figure under scene described in each.
Step 1221 is carried out according to following formula:
Wherein, m is that each organizes the sequence number of the first original depth-map of the first original image centering; The total quantity that the first original image that N gathers under referring to a scene is right, D
ijit is the pixel value of the first original depth-map;
refer to the pixel value of mean depth figure; v
i, jrepresent the variance referring to that pixel (i, j) is corresponding.
Step 1222 is carried out according to following formula:
Wherein, c
i, jrefer to the confidence level of pixel (i, j); v
i, jrefer to the variance of pixel (i, j);
E is natural constant, and λ is conversion parameter, and can control the size that variance is converted to confidence level, λ can be set to: 0.17.
In this step, Kinect device is used to gather many group colours and depth image pair at multiple visual angles of scene.For the image pair at each visual angle, can calculate an average color and depth image pair, composition average color and depth image are to group.In this, as foundation, calculate the variance depth information v often organizing image.
High variance yields correspond to low confidence, and vice versa.For each value v in variance depth information
i, japplication sigmoid function, calculates the degree of confidence c of their correspondences
i, j.
Step 11 comprises:
Step 111, gathers the first original image pair under a scene;
Step 112, when described first original image is to when comprising N frame the first original depth-map, according to described N frame first original depth-map, calculates the first mean depth figure; Average again after cumulative to the pixel value of all pixels in described N frame first mean depth figure, obtain the first pixel value D;
Step 113, when described first original image is to when comprising N+1 frame the first original depth-map, according to described N+1 frame first original depth-map, calculates the second mean depth figure; Average again after cumulative to the pixel value of all pixels in described N+1 frame second mean depth figure, obtain the second pixel value D';
Step 114, judges whether the absolute value of the difference between described first pixel value D and described second pixel value D' is less than threshold value;
Step 115, when the absolute value of the difference between described first pixel value D and described second pixel value D' is less than threshold value, stops gathering the first original image pair under this scenario.
Wherein, described first eigenvector comprise following one or more:
The radial distance of image slices vegetarian refreshments X (i, j)
The value of the following feature of image slices vegetarian refreshments X (i, j) and each described feature minimum value in the picture, maximal value, average and intermediate value;
Describedly to be characterized as:
The R red component of image slices vegetarian refreshments X (i, j), G green component, B blue component, D depth component;
The gradient information of pixel (i, j) in mean depth figure;
The marginal information of pixel (i, j) in mean depth figure;
The marginal information of pixel (i, j) in average color figure;
Pixel (i, j) carries out the filtered information in Gabor Garbo to average color figure;
Pixel (i, j) carries out the filtered information in Gabor Garbo to mean depth figure;
The texture information of pixel (i, j) in mean depth figure;
The texture information of pixel (i, j) in average color figure;
Pixel (i, j) arrives the distance of degree of depth hole in mean depth figure.
Wherein, described second feature vector comprise following one or more:
The radial distance of image slices vegetarian refreshments X (i, j)
The value of the following feature of image slices vegetarian refreshments X (i, j) and each described feature minimum value in the picture, maximal value, average and intermediate value;
Describedly to be characterized as:
The R red component of image slices vegetarian refreshments X (i, j), G green component, B blue component, D depth component;
The gradient information of pixel (i, j) in the second original depth-map;
The marginal information of pixel (i, j) in the second original depth-map;
The marginal information of pixel (i, j) in the second original color figure;
Pixel (i, j) carries out the filtered information in Gabor Garbo to the second original color figure;
Pixel (i, j) carries out the filtered information in Gabor Garbo to the second original depth-map;
The texture information of pixel (i, j) in the second original depth-map;
The texture information of pixel (i, j) in the second original color figure;
Pixel (i, j) arrives the distance of degree of depth hole in the second original depth-map.
The method extracting first eigenvector is similar with the Measures compare extracting second feature vector.There is concrete description hereinafter.
Claims (9)
1. the appraisal procedure of the confidence level of the degree of depth in depth map, it is characterized in that, described method comprises:
Obtain at least two group first original images pair under different scene, described first original image is to the first original color figure and the first original depth-map that comprise alignment;
According to described first original image pair, generation forecast device; Described fallout predictor represents the mapping relations between proper vector and degree of depth confidence level extracted from image;
Obtain the second original image pair of input, described second original image is to the second original color figure alignd under comprising a scene and the second original depth-map;
Extract the second feature vector that described second original image is right;
According to described fallout predictor and described second feature vector, predict the confidence level of the degree of depth of described second original depth-map.
2. method according to claim 1, is characterized in that, described according to described first original image pair, the step of generation forecast device comprises:
According at least two group first original images pair described under each scene, calculate average color figure corresponding to scene described in each and average depth map;
According to the described mean depth figure of scene described in each, generate the degree of depth confidence level figure under scene described in each;
According to the described average color figure under scene described in each and described mean depth figure, extract the feature that pixel is associated with degree of depth confidence level, composition first eigenvector
According to described first eigenvector
and the described degree of depth confidence level figure under scene described in each, utilize random forest method, the generation forecast device by training.
3. method according to claim 2, is characterized in that, described fallout predictor is:
Wherein, (i, j) represents the coordinate of pixel in described average color figure and described mean depth figure;
refer to the first eigenvector extracted from image;
F refers to the mapping relations between proper vector and degree of depth confidence level;
C
i, jrefer to the degree of depth confidence level of pixel (i, j).
4. method according to claim 2, is characterized in that, described according to the described mean depth figure under scene described in each, the step generating the degree of depth confidence level figure under scene described in each comprises:
According to the described mean depth figure under scene described in each, calculate the variance depth map under scene described in each;
Utilize the Sigmoid tangent bend function after adjustment, the described variance depth map under scene described in each is converted to the degree of depth confidence level figure under scene described in each.
5. method according to claim 4, is characterized in that, described according to the described mean depth figure under scene described in each, the step calculating the variance depth map under scene described in each is carried out according to following formula:
Wherein, m is that each organizes the sequence number of the first original depth-map of the first original image centering; The total quantity that the first original image that N gathers under referring to a scene is right, D
i,jit is the pixel value of the first original depth-map;
refer to the pixel value of mean depth figure; v
i, jrepresent the variance referring to that pixel (i, j) is corresponding.
6. method according to claim 4, it is characterized in that, Sigmoid tangent bend function after described utilization adjustment, the step of the degree of depth confidence level figure be converted under scene described in each by the described variance depth map under scene described in each is carried out according to following formula:
Wherein, c
i, jrefer to the confidence level of pixel (i, j); v
i, jrefer to the variance of pixel (i, j);
E is natural constant, and λ is conversion parameter.
7. method according to claim 1, is characterized in that, the step of at least two group first original images under described acquisition scene described in each comprises:
Gather the first original image pair under a scene;
When described first original image is to when comprising N frame the first original depth-map, according to described N frame first original depth-map, calculate the first mean depth figure; Average again after cumulative to the pixel value of all pixels in described N frame first mean depth figure, obtain the first pixel value D;
When described first original image is to when comprising N+1 frame the first original depth-map, according to described N+1 frame first original depth-map, calculate the second mean depth figure; Average again after cumulative to the pixel value of all pixels in described N+1 frame second mean depth figure, obtain the second pixel value D';
Judge whether the absolute value of the difference between described first pixel value D and described second pixel value D' is less than threshold value;
When the absolute value of the difference between described first pixel value D and described second pixel value D' is less than threshold value, stop gathering the first original image pair under this scenario.
8. method according to claim 2, is characterized in that, described first eigenvector comprise following one or more:
The radial distance of image slices vegetarian refreshments X (i, j)
The value of the following feature of image slices vegetarian refreshments X (i, j) and each described feature minimum value in the picture, maximal value, average and intermediate value;
Describedly to be characterized as:
The R red component of image slices vegetarian refreshments X (i, j), G green component, B blue component, D depth component;
The gradient information of pixel (i, j) in mean depth figure;
The marginal information of pixel (i, j) in mean depth figure;
The marginal information of pixel (i, j) in average color figure;
Pixel (i, j) carries out the filtered information in Gabor Garbo to average color figure;
Pixel (i, j) carries out the filtered information in Gabor Garbo to mean depth figure;
The texture information of pixel (i, j) in mean depth figure;
The texture information of pixel (i, j) in average color figure;
Pixel (i, j) arrives the distance of degree of depth hole in mean depth figure.
9. method according to claim 1, is characterized in that, described second feature vector comprise following one or more:
The radial distance of image slices vegetarian refreshments X (i, j)
The value of the following feature of image slices vegetarian refreshments X (i, j) and each described feature minimum value in the picture, maximal value, average and intermediate value;
Describedly to be characterized as:
The R red component of image slices vegetarian refreshments X (i, j), G green component, B blue component, D depth component;
The gradient information of pixel (i, j) in the second original depth-map;
The marginal information of pixel (i, j) in the second original depth-map;
The marginal information of pixel (i, j) in the second original color figure;
Pixel (i, j) carries out the filtered information in Gabor Garbo to the second original color figure;
Pixel (i, j) carries out the filtered information in Gabor Garbo to the second original depth-map;
The texture information of pixel (i, j) in the second original depth-map;
The texture information of pixel (i, j) in the second original color figure;
Pixel (i, j) arrives the distance of degree of depth hole in the second original depth-map.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510548522.XA CN105139401A (en) | 2015-08-31 | 2015-08-31 | Depth credibility assessment method for depth map |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510548522.XA CN105139401A (en) | 2015-08-31 | 2015-08-31 | Depth credibility assessment method for depth map |
Publications (1)
Publication Number | Publication Date |
---|---|
CN105139401A true CN105139401A (en) | 2015-12-09 |
Family
ID=54724733
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510548522.XA Pending CN105139401A (en) | 2015-08-31 | 2015-08-31 | Depth credibility assessment method for depth map |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN105139401A (en) |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106780543A (en) * | 2017-01-13 | 2017-05-31 | 深圳市唯特视科技有限公司 | A kind of double framework estimating depths and movement technique based on convolutional neural networks |
CN108564620A (en) * | 2018-03-27 | 2018-09-21 | 中国人民解放军国防科技大学 | Scene depth estimation method for light field array camera |
TWI641264B (en) * | 2017-03-30 | 2018-11-11 | 晶睿通訊股份有限公司 | Image processing system and lens state determination method |
CN109146886A (en) * | 2018-08-19 | 2019-01-04 | 沈阳农业大学 | A kind of RGBD image, semantic segmentation optimization method based on depth density |
CN109767467A (en) * | 2019-01-22 | 2019-05-17 | Oppo广东移动通信有限公司 | Image processing method, device, electronic equipment and computer readable storage medium |
CN110400340A (en) * | 2019-07-11 | 2019-11-01 | Oppo广东移动通信有限公司 | Depth map treating method and apparatus |
CN110400273A (en) * | 2019-07-11 | 2019-11-01 | Oppo广东移动通信有限公司 | Filtering method, device, electronic equipment and the readable storage medium storing program for executing of depth data |
WO2023245378A1 (en) * | 2022-06-20 | 2023-12-28 | 北京小米移动软件有限公司 | Method and apparatus for determining confidence level of depth information of image, and storage medium |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102609941A (en) * | 2012-01-31 | 2012-07-25 | 北京航空航天大学 | Three-dimensional registering method based on ToF (Time-of-Flight) depth camera |
WO2014065887A1 (en) * | 2012-10-24 | 2014-05-01 | Lsi Corporation | Image processing method and apparatus for elimination of depth artifacts |
CN104166987A (en) * | 2014-07-08 | 2014-11-26 | 哈尔滨师范大学 | Parallax estimation method based on improved adaptive weighted summation and belief propagation |
CN104574375A (en) * | 2014-12-23 | 2015-04-29 | 浙江大学 | Image significance detection method combining color and depth information |
-
2015
- 2015-08-31 CN CN201510548522.XA patent/CN105139401A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102609941A (en) * | 2012-01-31 | 2012-07-25 | 北京航空航天大学 | Three-dimensional registering method based on ToF (Time-of-Flight) depth camera |
WO2014065887A1 (en) * | 2012-10-24 | 2014-05-01 | Lsi Corporation | Image processing method and apparatus for elimination of depth artifacts |
CN104166987A (en) * | 2014-07-08 | 2014-11-26 | 哈尔滨师范大学 | Parallax estimation method based on improved adaptive weighted summation and belief propagation |
CN104574375A (en) * | 2014-12-23 | 2015-04-29 | 浙江大学 | Image significance detection method combining color and depth information |
Non-Patent Citations (2)
Title |
---|
DOBOS REYNOLDS 等: "Capturing time-of-flight data with confidence", 《THE PROCEEDINGS OF 2011 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION》 * |
XIBIN SONG 等: "Estimation of Kinect depth confidence through self-training", 《THE VISUAL COMPUTER》 * |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106780543A (en) * | 2017-01-13 | 2017-05-31 | 深圳市唯特视科技有限公司 | A kind of double framework estimating depths and movement technique based on convolutional neural networks |
TWI641264B (en) * | 2017-03-30 | 2018-11-11 | 晶睿通訊股份有限公司 | Image processing system and lens state determination method |
US10573014B2 (en) | 2017-03-30 | 2020-02-25 | Vivotek Inc. | Image processing system and lens state determination method |
CN108564620A (en) * | 2018-03-27 | 2018-09-21 | 中国人民解放军国防科技大学 | Scene depth estimation method for light field array camera |
CN109146886A (en) * | 2018-08-19 | 2019-01-04 | 沈阳农业大学 | A kind of RGBD image, semantic segmentation optimization method based on depth density |
CN109767467A (en) * | 2019-01-22 | 2019-05-17 | Oppo广东移动通信有限公司 | Image processing method, device, electronic equipment and computer readable storage medium |
CN109767467B (en) * | 2019-01-22 | 2020-11-10 | Oppo广东移动通信有限公司 | Image processing method, image processing device, electronic equipment and computer readable storage medium |
CN110400340A (en) * | 2019-07-11 | 2019-11-01 | Oppo广东移动通信有限公司 | Depth map treating method and apparatus |
CN110400273A (en) * | 2019-07-11 | 2019-11-01 | Oppo广东移动通信有限公司 | Filtering method, device, electronic equipment and the readable storage medium storing program for executing of depth data |
CN110400340B (en) * | 2019-07-11 | 2022-03-22 | Oppo广东移动通信有限公司 | Depth map processing method and device |
CN110400273B (en) * | 2019-07-11 | 2022-03-22 | Oppo广东移动通信有限公司 | Depth data filtering method and device, electronic equipment and readable storage medium |
WO2023245378A1 (en) * | 2022-06-20 | 2023-12-28 | 北京小米移动软件有限公司 | Method and apparatus for determining confidence level of depth information of image, and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN105139401A (en) | Depth credibility assessment method for depth map | |
CN108537191B (en) | Three-dimensional face recognition method based on structured light camera | |
CN103996192B (en) | Non-reference image quality evaluation method based on high-quality natural image statistical magnitude model | |
CN113936139B (en) | Scene aerial view reconstruction method and system combining visual depth information and semantic segmentation | |
CN102982334B (en) | The sparse disparities acquisition methods of based target edge feature and grey similarity | |
CN101996407B (en) | Colour calibration method for multiple cameras | |
CN106096561A (en) | Infrared pedestrian detection method based on image block degree of depth learning characteristic | |
CN107560592B (en) | Precise distance measurement method for photoelectric tracker linkage target | |
CN104318569A (en) | Space salient region extraction method based on depth variation model | |
CN103927758B (en) | Saliency detection method based on contrast ratio and minimum convex hull of angular point | |
CN101610425B (en) | Method for evaluating stereo image quality and device | |
CN107392963A (en) | A kind of imitative hawkeye moving target localization method for soft autonomous air refuelling | |
DeWitt et al. | Creating high-resolution bare-earth digital elevation models (DEMs) from stereo imagery in an area of densely vegetated deciduous forest using combinations of procedures designed for lidar point cloud filtering | |
CN103198475B (en) | Based on the total focus synthetic aperture perspective imaging method that multilevel iteration visualization is optimized | |
CN105139355A (en) | Method for enhancing depth images | |
CN106355607B (en) | A kind of width baseline color image template matching method | |
CN104517095A (en) | Head division method based on depth image | |
CN113050074B (en) | Camera and laser radar calibration system and calibration method in unmanned environment perception | |
CN101246595A (en) | Multi-view point data splitting method of optical three-dimensional scanning system | |
US20140125667A1 (en) | Roof Generation And Texturing Of 3D Models | |
CN109308451A (en) | A kind of high score data information extraction system and method | |
CN101908153A (en) | Method for estimating head postures in low-resolution image treatment | |
CN103281513A (en) | Method for recognizing pedestrian in monitoring system without overlapped area | |
CN116071424A (en) | Fruit space coordinate positioning method based on monocular vision | |
CN109410175A (en) | SAR radar imagery quality quick automatic evaluation method based on multiple subarea images match |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20151209 |
|
RJ01 | Rejection of invention patent application after publication |