CN114549297A - Unsupervised monocular depth estimation method based on uncertain analysis - Google Patents
Unsupervised monocular depth estimation method based on uncertain analysis Download PDFInfo
- Publication number
- CN114549297A CN114549297A CN202111185472.5A CN202111185472A CN114549297A CN 114549297 A CN114549297 A CN 114549297A CN 202111185472 A CN202111185472 A CN 202111185472A CN 114549297 A CN114549297 A CN 114549297A
- Authority
- CN
- China
- Prior art keywords
- depth
- depth estimation
- uncertainty
- function
- unsupervised
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 33
- 238000004458 analytical method Methods 0.000 title claims description 7
- 230000006870 function Effects 0.000 claims abstract description 48
- 238000013076 uncertainty analysis Methods 0.000 claims description 13
- 238000012549 training Methods 0.000 claims description 12
- 238000005457 optimization Methods 0.000 claims description 10
- 238000005286 illumination Methods 0.000 claims description 8
- 238000006243 chemical reaction Methods 0.000 claims description 7
- 230000008569 process Effects 0.000 claims description 7
- 239000000126 substance Substances 0.000 claims 2
- 238000013527 convolutional neural network Methods 0.000 abstract description 2
- 238000013135 deep learning Methods 0.000 abstract description 2
- 238000002474 experimental method Methods 0.000 description 8
- 230000000694 effects Effects 0.000 description 7
- 238000010586 diagram Methods 0.000 description 3
- 238000002679 ablation Methods 0.000 description 2
- 238000013528 artificial neural network Methods 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 230000006835 compression Effects 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- HPTJABJPZMULFH-UHFFFAOYSA-N 12-[(Cyclohexylcarbamoyl)amino]dodecanoic acid Chemical compound OC(=O)CCCCCCCCCCCNC(=O)NC1CCCCC1 HPTJABJPZMULFH-UHFFFAOYSA-N 0.000 description 1
- 241000282412 Homo Species 0.000 description 1
- 230000002159 abnormal effect Effects 0.000 description 1
- 230000003190 augmentative effect Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 230000002950 deficient Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000011478 gradient descent method Methods 0.000 description 1
- 125000001475 halogen functional group Chemical group 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 238000010606 normalization Methods 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 238000011084 recovery Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4007—Scaling of whole images or parts thereof, e.g. expanding or contracting based on interpolation, e.g. bilinear interpolation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/90—Determination of colour characteristics
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Image Processing (AREA)
Abstract
The invention discloses an unsupervised monocular depth estimation method based on uncertainty, which firstly provides an unsupervised depth estimation network based on uncertainty for improving the problem of low prediction depth precision in monocular depth estimation, and the uncertainty learning method solves the problem of strong expression capability of a convolutional neural network used for monocular depth estimation at present. The method of the invention trains the deep learning network in an unsupervised mode, avoids the limitation of a deep label, and the loss function of the invention consists of three parts, namely brightness loss, smoothness loss and uncertainty loss, not only can estimate the depth, but also can obtain the confidence coefficient of the estimated depth through predicting the variance.
Description
Technical Field
The invention relates to the technical field of computer vision, in particular to an unsupervised monocular depth estimation method based on uncertain analysis.
Background
Depth estimation is often critical to various advanced tasks in computer vision, such as autopilot, augmented reality, and the like. The method is also a core technology in an intelligent auxiliary driving system and an intelligent vehicle-mounted vision system of an automobile. By combining the depth information of depth estimation, the running state of the vehicle can be monitored, and a forward collision early warning system is perfected. Generally, the vehicle runs faster, so the requirement for vehicle distance measurement is higher, and the cooperativity between cameras in a binocular system and a multi-view system is not well adjusted, so the invention is developed based on a monocular camera. Monocular vision is more cost effective than binocular and multi-ocular systems from a cost perspective. In addition, the calculated amount of monocular vision in the aspect of data processing is much smaller than that of binocular vision and multi-ocular vision, real-time performance can be better considered, and the method is a research hotspot and frontier field in modern intelligent auxiliary driving systems and intelligent vehicle-mounted vision systems. At present, monocular depth data sets are deficient, and training can be performed only in scenes under specific data sets. In addition, details and depths around a depth map object are unclear in monocular depth estimation, dynamic objects interfere more, and the reliability of an output result cannot be evaluated by a convolutional neural network, which is also an urgent problem to be solved. Based on the problems, the invention provides an unsupervised monocular depth estimation method based on uncertainty.
Disclosure of Invention
In order to overcome the defects in the prior art, the embodiment of the invention provides an unsupervised monocular depth estimation method based on uncertain analysis, which is used for solving the problems in the background art.
The invention discloses: an unsupervised monocular depth estimation method based on uncertain analysis is characterized by comprising the following steps of:
step 1: an unsupervised depth estimation network based on uncertainty is provided, and the depth prediction precision in monocular depth estimation is improved;
step 2: based on the step 1, the confidence coefficient of the estimated depth is predicted through modeling uncertainty, meanwhile, the model prediction precision is improved, and the uncertainty of an output result is quantified;
and step 3: based on the step 1-2, a Retinex illumination theory is used for constructing a brightness loss function, and the interference caused by dynamic objects in the scene is solved through operation conversion according to the basic theory of a Retinex algorithm.
Preferably, the step 1 specifically comprises:
first, a likelihood function is defined:
p(y|y*)=N(y*,σ2);
wherein y denotes the depth of observation, y*Representing the depth, σ, of the model output2Representing an observed noise variance;
secondly, solving the likelihood function:
again, an objective function is established:
depth estimation of images is a regression task, L1Is superior to L2A penalty function;
then, the loss function for the uncertainty analysis is as follows:
finally, the resulting loss function is:
L=LR+LS+LU。
preferably, two consecutive frames I sampled from a given video that is unmarkedtAnd It-1First, estimate its depth map D using a depth networktAnd Dt-1;
Then using the pose network PabTraining relative 6D poses between cameras; depth map D with predictiontAnd relative camera pose PabBy differentiable bilinear interpolationt-1Is formed by deformationSimilarly, an image is obtained
Finally, willInput into a depth net to obtainBy means of uncertainty analysis, inAndform a loss function L therebetweenU。
Preferably, for the depth estimation network, the invention improves on DispNet, which takes a single RGB image as input and outputs a depth map.
Preferably, for the posture network, the invention uses a network DispNet without mask predicted branches.
Preferably, the step 3 specifically comprises:
firstly, according to the basic theory of Retinex algorithm, the following expression is obtained:
I(x,y)=R(x,y)×L(x,y);
secondly, the process of solving the incident component by convolution operation with a low-pass filter can be expressed as:
L(x,y)=I(x,y)*G(x,y);
thirdly, the mathematical solving process of the single-scale Retinex algorithm is shown as follows:
wherein i belongs to R, G and B and represents R, G and B color channels, Ri(x, y) represents the pixel value of the reflected image in the ith color channel, Ii(x, y) represents the pixel value of the original image I at the ith color channel (x, y), which represents a gaussian convolution operation, and G (x, y) represents a gaussian surround function;
where σ represents the standard deviation in a gaussian function, referred to herein as the scale function;
then, the depth estimation is a regression task, and the most commonly used loss functions in the optimization of the regression task comprise an L2 loss function and an L1 loss function;
finally, from the above, the conversion formula yields:
where N represents a pixel in the image, ri(x, y) represents incident light of an image.
Preferably, smoothness is included before regularizing the estimated depth map, and an edge-aware smoothness penalty is used, which is given by the formula:
the invention has the following beneficial effects: the invention provides an uncertainty-based monocular depth estimation method, which solves the problem that the detail depth around a depth map object is unclear, and the most advanced performance can be obtained on a KITTI data set.
In order to make the aforementioned and other objects, features and advantages of the invention comprehensible, preferred embodiments accompanied with figures are described in detail below.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts.
FIG. 1 is a basic diagram of the Retinex algorithm of the present invention;
FIG. 2 is a schematic diagram of an unsupervised monocular depth estimation method of the present invention;
FIG. 3 is a schematic diagram of the DispNet network structure of the present invention;
FIG. 4 is a graph I of the experimental results of the present invention;
FIG. 5 is a second graph showing the results of the experiment according to the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
An unsupervised monocular depth estimation method based on uncertain analysis comprises the following steps:
step 1: an unsupervised depth estimation network based on uncertainty is provided, and the depth prediction precision in monocular depth estimation is improved;
step 2: based on the step 1, the confidence coefficient of the estimated depth is predicted through modeling uncertainty, meanwhile, the model prediction precision is improved, and the uncertainty of an output result is quantified;
and step 3: based on the step 1-2, a Retinex illumination theory is used for constructing a brightness loss function, and the interference of dynamic objects in the scene is solved through operation and conversion according to the basic theory of a Retinex algorithm.
Further, the step 1 specifically comprises: the uncertainty in neural networks is largely divided into two areas: model uncertainty and random uncertainty; the model uncertainty mainly refers to the uncertainty of model parameters, and when a plurality of models have good effects, the final model parameters need to be selected from the models; when the amount of input data is large enough, the model uncertainty becomes small; in general, training data is enough during training, so random uncertainty in uncertainty analysis accounts for a main part;
first, a likelihood function is defined:
p(y|y*)=N(y*,σ2);
wherein y denotes the depth of observation, y*Representing the depth, σ, of the model output2Representing an observed noise variance;
secondly, solving the likelihood function:
then, an objective function is established:
depth estimation of images is a regression task, L1Is superior to L2A penalty function; because L is1The optimization effect on small prediction errors is good, and the optimization just accords with the characteristic of depth estimation; therefore, the loss function for the uncertainty analysis is as follows:
finally, the resulting loss function is:
L=LR+LS+LU。
further, referring to FIG. 2, two consecutive frames I sampled from a given unmarked videotAnd It-1First, estimate its depth map D using a depth networktAnd Dt-1Then using the pose network PabTraining relative 6D poses between cameras; depth map D with predictiontAnd relative camera pose PabThe invention uses differentiable bilinear interpolation to convert It-1Is transformed to synthesizeSimilarly, an image is obtainedFinally, willInput into a depth net to obtainBy means of uncertainty analysis, inAndform a loss function L therebetweenU;
Further, for the depth estimation network, the present invention improves on DispNet, which takes a single RGB image as input and outputs a depth map.
Further, referring to fig. 3, for the posture network, the present invention uses a network DispNet without mask predicted branches.
Further, the step 3 specifically includes: the algorithm of Retinex is based on three assumptions:
(1) the real physical world is colorless and the colors seen by humans are the result of the interaction of natural light with matter in the objective world.
(2) Each color range in the object is composed of red, green, and blue of a certain wavelength.
(3) The color of each region is determined by red, green and blue.
Referring to fig. 1, a Retinex algorithm is different from a linear and nonlinear method that only can enhance a certain feature of an image, Retinex can achieve balance in three aspects of compression, edge enhancement and color normalization in a dynamic range, so that different types of images can be adaptively enhanced, and color fidelity, image edge portion enhancement and dynamic range compression can be obviously improved.
Through analyzing the illumination component and the reflection component, people can judge the illumination information and the reflection information in the image, thereby solving the problem of image brightness and color change caused by illumination change, particularly, the illumination information influencing the vision of people is removed through various conversion methods, the reflection information of an object is retained to the maximum extent, and because the attribute information of the object is contained in the reflection component, a human visual system imitating the human visual system develops a Retinex algorithm all the time, the Retinex algorithm is improved from a single-scale Retinex algorithm to a multiscale weighted average Retinex algorithm, and then the Retinex algorithm is developed into a color recovery multiscale Retinex algorithm;
firstly, according to the basic theory of Retinex algorithm, the following expression is obtained:
I(x,y)=R(x,y)×L(x,y);
secondly, the main principle of the single-scale Retinex algorithm is to perform convolution operation on three channels of the image and a central surrounding function respectively, and the image after the convolution operation is regarded as the estimation of the illumination component of the original image;
the process of solving for the incident component by convolution operation with a low pass filter can be expressed as:
L(x,y)=I(x,y)*G(x,y);
thirdly, the mathematical solving process of the single-scale Retinex algorithm is shown as follows:
wherein i belongs to R, G and B and represents R, G and B color channels, Ri(x, y) represents the pixel value of the reflected image in the ith color channel, Ii(x, y) represents the pixel value of the original image I at the ith color channel (x, y), which represents a gaussian convolution operation, and G (x, y) represents a gaussian surround function;
where σ represents the standard deviation in a gaussian function, referred to herein as the scale function; the size of the standard deviation has great influence on the Retinex algorithm; the smaller the sigma is, the better detail information of the enhanced image can be obtained, but image distortion and halo phenomena are easy to occur; the larger the sigma is, the better the color of the enhanced image is kept, but the image has a larger sharpening degree and a poorer contrast enhancement effect;
then, the depth estimation is a regression task, and the most commonly used loss functions in the optimization of the regression task comprise an L2 loss function and an L1 loss function;
the square operation makes the L2 loss function sensitive to abnormal values, has good optimization effect on large prediction errors, and has poor capability of further optimization on small prediction errors; the L1 loss function has a good optimization effect on small prediction errors, the optimization effect on large prediction errors is general, and the effect of the L1 loss function is slightly excellent in actual training; the uncertainty Loss function provided by the invention combines L1 Loss and heteroscedastic random uncertainty in a neural network, and consists of a residual regression term and an uncertainty regularization term;
finally, from the above, the conversion formula yields:
where N represents a pixel in the image, ri(x, y) represents incident light of an image.
Further, because luminosity loss does not provide sufficient information in low-texture or uniform regions of a scene, existing work incorporates smoothness prior to regularizing the estimated depth map; the edge-perceived smoothness loss used is given by:
the invention carries out the experiment of monocular depth estimation, which comprises the following steps: the experimental environment provided by the invention comprises a software environment and a hardware environment, wherein the software environment comprises: windows 1064-bit operating system, CUDA 9.1, cuDNN 7.1, Pythrch deep learning framework, Python 3.7.0 and MATLAB R2018 a; the hardware environment is as follows: intel (R) core (TM) i7-7700 CPU @3.60GHz processor, 32GB RAM and NVIDIA GeForce GTX 1080Ti GPU, 11 GB;
according to the invention, a KITTI data set is adopted to carry out monocular depth estimation experiments, about 42000 pictures are included in a training set, and no ground true value exists; in the training process, a random gradient descent method is adopted for optimization solution, the weight parameters of the training network are continuously updated by using a back propagation algorithm, the initial learning rate is set to be 0.001, the momentum factor is set to be 0.9, and the weight attenuation factor is set to be 0.0005; the learning rate is closely related to the convergence speed of the training network, the network model cannot be converged if the learning rate is too large, and the convergence speed of the network model becomes slow if the learning rate is too small; in the invention, the maximum iteration number of the training network is 20000, the learning rate of the previous 12000 times is set to be 0.001, the learning rate of the previous 12000 times to 16000 times is set to be 0.0001, and the learning rate of the previous 16000 times is set to be 0.00001;
in order to objectively evaluate the proposed monocular depth estimation model, the model is quantitatively analyzed using the following four criteria:
mean absolute relative error (Rel):
root mean square error: (RMS):
mean log10 error (RMSlog):
accuracy within the threshold range:
referring to fig. 4, through the unsupervised depth estimation method provided by the invention, experiments are carried out on random pictures in a test set, the topmost picture in fig. 4 is an original picture, the bottommost picture is the algorithm provided by the invention, and at the position of a dotted line frame in fig. 4, the phenomenon that the depth of the algorithm provided by the invention does not excessively drift around traffic lights is obviously seen, the precision is improved, and the reason of dynamic object interference is also solved;
referring to fig. 5, four pictures of adjacent frames are also realized by the algorithm provided by the present invention, and it can be easily seen that the depth around the traffic light is well improved.
Referring to tables 1 and 2, to further verify the utility of the present invention, the following are respectively: three methods of basic network, basic network and brightness loss, basic network and uncertainty, basic network and brightness loss and uncertainty are used for carrying out experiments, and the three experimental methods have the same smoothness loss; the table data shows that the problem of low precision is solved by the brightness loss and the uncertainty loss in the invention;
TABLE 1 ablation experiment 1 (resolution of input image: 416X128)
TABLE 2 ablation experiment 2 (resolution of input image: 832x256)
The principle and the implementation mode of the invention are explained by applying specific embodiments in the invention, and the description of the embodiments is only used for helping to understand the method and the core idea of the invention; meanwhile, for a person skilled in the art, according to the idea of the present invention, there may be variations in the specific embodiments and the application scope, and in summary, the content of the present specification should not be construed as a limitation to the present invention.
Claims (7)
1. An unsupervised monocular depth estimation method based on uncertain analysis is characterized by comprising the following steps of:
step 1: an unsupervised depth estimation network based on uncertainty is provided, and the depth prediction precision in monocular depth estimation is improved;
step 2: based on the step 1, the confidence coefficient of the estimated depth is predicted through modeling uncertainty, meanwhile, the model prediction precision is improved, and the uncertainty of an output result is quantized;
and step 3: based on the step 1-2, a Retinex illumination theory is used for constructing a brightness loss function, and the interference of dynamic objects in the scene is solved through operation and conversion according to the basic theory of a Retinex algorithm.
2. The unsupervised monocular depth estimation method based on uncertainty analysis of claim 1, further comprising:
the step 1 specifically comprises the following steps:
first, a likelihood function is defined:
whereinWhich represents the depth of the observation(s),the depth of the output of the model is represented,representing an observed noise variance;
secondly, solving the likelihood function:
again, an objective function is established:
then, the loss function for the uncertainty analysis is as follows:
finally, the resulting loss function is:
3. the unsupervised monocular depth estimation method based on uncertainty analysis of claim 2, further comprising: two consecutive frames sampled from a given unmarked videoAndfirst, estimate its depth map using a depth networkAnd;
then using the attitude networkTraining relative 6D poses between the cameras; depth map with predictionAnd relative camera poseBy differentiable bilinear interpolationIs formed by deformationSimilarly, an image is obtained;
4. The unsupervised monocular depth estimation method based on uncertainty analysis of claim 1, further comprising: for the depth estimation network, the present invention improves on DispNet, which takes a single RGB image as input and outputs a depth map.
5. The unsupervised monocular depth estimation method based on uncertainty analysis of claim 3, further comprising: for the posture network, the invention uses the network DispNet without mask predicted branches.
6. The unsupervised monocular depth estimation method based on uncertainty analysis of claim 1, further comprising: the step 3 specifically comprises the following steps:
firstly, according to the basic theory of Retinex algorithm, the following expression is obtained:
secondly, the process of solving the incident component by convolution operation with a low-pass filter can be expressed as:
thirdly, the mathematical solving process of the single-scale Retinex algorithm is shown as follows:
wherein the content of the first and second substances,indicating the color channels of R, G, B,indicating a reflected image is onThe pixel values of the individual color channels,representing an original imageIn the first placeColor channelThe pixel values at, represent a gaussian convolution operation,representing a gaussian surround function;
wherein the content of the first and second substances,represents the standard deviation in a gaussian function, referred to herein as a scale function;
then, the depth estimation is a regression task, and the most commonly used loss functions in the optimization of the regression task comprise an L2 loss function and an L1 loss function;
finally, from the above, the conversion formula yields:
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111185472.5A CN114549297A (en) | 2021-10-12 | 2021-10-12 | Unsupervised monocular depth estimation method based on uncertain analysis |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111185472.5A CN114549297A (en) | 2021-10-12 | 2021-10-12 | Unsupervised monocular depth estimation method based on uncertain analysis |
Publications (1)
Publication Number | Publication Date |
---|---|
CN114549297A true CN114549297A (en) | 2022-05-27 |
Family
ID=81668506
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202111185472.5A Pending CN114549297A (en) | 2021-10-12 | 2021-10-12 | Unsupervised monocular depth estimation method based on uncertain analysis |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114549297A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114782782A (en) * | 2022-06-20 | 2022-07-22 | 武汉大学 | Uncertainty quantification method for learning performance of monocular depth estimation model |
CN114820755A (en) * | 2022-06-24 | 2022-07-29 | 武汉图科智能科技有限公司 | Depth map estimation method and system |
-
2021
- 2021-10-12 CN CN202111185472.5A patent/CN114549297A/en active Pending
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114782782A (en) * | 2022-06-20 | 2022-07-22 | 武汉大学 | Uncertainty quantification method for learning performance of monocular depth estimation model |
CN114782782B (en) * | 2022-06-20 | 2022-10-04 | 武汉大学 | Uncertainty quantification method for learning performance of monocular depth estimation model |
CN114820755A (en) * | 2022-06-24 | 2022-07-29 | 武汉图科智能科技有限公司 | Depth map estimation method and system |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Li et al. | Image dehazing using residual-based deep CNN | |
CN108986050B (en) | Image and video enhancement method based on multi-branch convolutional neural network | |
CN109685072B (en) | Composite degraded image high-quality reconstruction method based on generation countermeasure network | |
Negru et al. | Exponential contrast restoration in fog conditions for driving assistance | |
CN103914813B (en) | The restored method of colored haze image defogging and illumination compensation | |
KR102261532B1 (en) | Method and system for image dehazing using single scale image fusion | |
CN110544213A (en) | Image defogging method based on global and local feature fusion | |
CN109389569B (en) | Monitoring video real-time defogging method based on improved DehazeNet | |
CN114549297A (en) | Unsupervised monocular depth estimation method based on uncertain analysis | |
CN115063329A (en) | Visible light and infrared image fusion enhancement method and system under low-illumination environment | |
CN116797488A (en) | Low-illumination image enhancement method based on feature fusion and attention embedding | |
Agrawal et al. | A comprehensive review on analysis and implementation of recent image dehazing methods | |
CN111476213A (en) | Method and device for filling covering area of shelter based on road image | |
CN112703532B (en) | Image processing method, device, equipment and storage medium | |
CN110969171A (en) | Image classification model, method and application based on improved convolutional neural network | |
CN113284061B (en) | Underwater image enhancement method based on gradient network | |
CN114170286B (en) | Monocular depth estimation method based on unsupervised deep learning | |
CN115775376A (en) | Crowd counting method based on low-light image enhancement | |
CN106657948A (en) | low illumination level Bayer image enhancing method and enhancing device | |
CN116596792B (en) | Inland river foggy scene recovery method, system and equipment for intelligent ship | |
Moghimi et al. | Real-time underwater image resolution enhancement using super-resolution with deep convolutional neural networks | |
CN110223240A (en) | Image defogging method, system and storage medium based on color decaying priori | |
CN115131229A (en) | Image noise reduction and filtering data processing method and device and computer equipment | |
Liang et al. | Learning to remove sandstorm for image enhancement | |
CN109064444B (en) | Track slab disease detection method based on significance analysis |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |