CN114549297A - Unsupervised monocular depth estimation method based on uncertain analysis - Google Patents

Unsupervised monocular depth estimation method based on uncertain analysis Download PDF

Info

Publication number
CN114549297A
CN114549297A CN202111185472.5A CN202111185472A CN114549297A CN 114549297 A CN114549297 A CN 114549297A CN 202111185472 A CN202111185472 A CN 202111185472A CN 114549297 A CN114549297 A CN 114549297A
Authority
CN
China
Prior art keywords
depth
depth estimation
uncertainty
function
unsupervised
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111185472.5A
Other languages
Chinese (zh)
Inventor
宋传学
齐春阳
彭思仑
宋世欣
肖峰
王达
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jilin University
Original Assignee
Jilin University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jilin University filed Critical Jilin University
Priority to CN202111185472.5A priority Critical patent/CN114549297A/en
Publication of CN114549297A publication Critical patent/CN114549297A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4007Scaling of whole images or parts thereof, e.g. expanding or contracting based on interpolation, e.g. bilinear interpolation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/90Determination of colour characteristics
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses an unsupervised monocular depth estimation method based on uncertainty, which firstly provides an unsupervised depth estimation network based on uncertainty for improving the problem of low prediction depth precision in monocular depth estimation, and the uncertainty learning method solves the problem of strong expression capability of a convolutional neural network used for monocular depth estimation at present. The method of the invention trains the deep learning network in an unsupervised mode, avoids the limitation of a deep label, and the loss function of the invention consists of three parts, namely brightness loss, smoothness loss and uncertainty loss, not only can estimate the depth, but also can obtain the confidence coefficient of the estimated depth through predicting the variance.

Description

Unsupervised monocular depth estimation method based on uncertain analysis
Technical Field
The invention relates to the technical field of computer vision, in particular to an unsupervised monocular depth estimation method based on uncertain analysis.
Background
Depth estimation is often critical to various advanced tasks in computer vision, such as autopilot, augmented reality, and the like. The method is also a core technology in an intelligent auxiliary driving system and an intelligent vehicle-mounted vision system of an automobile. By combining the depth information of depth estimation, the running state of the vehicle can be monitored, and a forward collision early warning system is perfected. Generally, the vehicle runs faster, so the requirement for vehicle distance measurement is higher, and the cooperativity between cameras in a binocular system and a multi-view system is not well adjusted, so the invention is developed based on a monocular camera. Monocular vision is more cost effective than binocular and multi-ocular systems from a cost perspective. In addition, the calculated amount of monocular vision in the aspect of data processing is much smaller than that of binocular vision and multi-ocular vision, real-time performance can be better considered, and the method is a research hotspot and frontier field in modern intelligent auxiliary driving systems and intelligent vehicle-mounted vision systems. At present, monocular depth data sets are deficient, and training can be performed only in scenes under specific data sets. In addition, details and depths around a depth map object are unclear in monocular depth estimation, dynamic objects interfere more, and the reliability of an output result cannot be evaluated by a convolutional neural network, which is also an urgent problem to be solved. Based on the problems, the invention provides an unsupervised monocular depth estimation method based on uncertainty.
Disclosure of Invention
In order to overcome the defects in the prior art, the embodiment of the invention provides an unsupervised monocular depth estimation method based on uncertain analysis, which is used for solving the problems in the background art.
The invention discloses: an unsupervised monocular depth estimation method based on uncertain analysis is characterized by comprising the following steps of:
step 1: an unsupervised depth estimation network based on uncertainty is provided, and the depth prediction precision in monocular depth estimation is improved;
step 2: based on the step 1, the confidence coefficient of the estimated depth is predicted through modeling uncertainty, meanwhile, the model prediction precision is improved, and the uncertainty of an output result is quantified;
and step 3: based on the step 1-2, a Retinex illumination theory is used for constructing a brightness loss function, and the interference caused by dynamic objects in the scene is solved through operation conversion according to the basic theory of a Retinex algorithm.
Preferably, the step 1 specifically comprises:
first, a likelihood function is defined:
p(y|y*)=N(y*2);
wherein y denotes the depth of observation, y*Representing the depth, σ, of the model output2Representing an observed noise variance;
secondly, solving the likelihood function:
Figure RE-GDA0003577100850000021
again, an objective function is established:
Figure RE-GDA0003577100850000022
depth estimation of images is a regression task, L1Is superior to L2A penalty function;
then, the loss function for the uncertainty analysis is as follows:
Figure RE-GDA0003577100850000023
finally, the resulting loss function is:
L=LR+LS+LU
preferably, two consecutive frames I sampled from a given video that is unmarkedtAnd It-1First, estimate its depth map D using a depth networktAnd Dt-1
Then using the pose network PabTraining relative 6D poses between cameras; depth map D with predictiontAnd relative camera pose PabBy differentiable bilinear interpolationt-1Is formed by deformation
Figure RE-GDA0003577100850000024
Similarly, an image is obtained
Figure RE-GDA0003577100850000025
Finally, will
Figure RE-GDA0003577100850000026
Input into a depth net to obtain
Figure RE-GDA0003577100850000027
By means of uncertainty analysis, in
Figure RE-GDA0003577100850000028
And
Figure RE-GDA0003577100850000029
form a loss function L therebetweenU
Preferably, for the depth estimation network, the invention improves on DispNet, which takes a single RGB image as input and outputs a depth map.
Preferably, for the posture network, the invention uses a network DispNet without mask predicted branches.
Preferably, the step 3 specifically comprises:
firstly, according to the basic theory of Retinex algorithm, the following expression is obtained:
I(x,y)=R(x,y)×L(x,y);
secondly, the process of solving the incident component by convolution operation with a low-pass filter can be expressed as:
L(x,y)=I(x,y)*G(x,y);
thirdly, the mathematical solving process of the single-scale Retinex algorithm is shown as follows:
Figure RE-GDA0003577100850000031
wherein i belongs to R, G and B and represents R, G and B color channels, Ri(x, y) represents the pixel value of the reflected image in the ith color channel, Ii(x, y) represents the pixel value of the original image I at the ith color channel (x, y), which represents a gaussian convolution operation, and G (x, y) represents a gaussian surround function;
Figure RE-GDA0003577100850000032
where σ represents the standard deviation in a gaussian function, referred to herein as the scale function;
then, the depth estimation is a regression task, and the most commonly used loss functions in the optimization of the regression task comprise an L2 loss function and an L1 loss function;
Figure RE-GDA0003577100850000033
Figure RE-GDA0003577100850000034
finally, from the above, the conversion formula yields:
Figure RE-GDA0003577100850000035
Figure RE-GDA0003577100850000036
where N represents a pixel in the image, ri(x, y) represents incident light of an image.
Preferably, smoothness is included before regularizing the estimated depth map, and an edge-aware smoothness penalty is used, which is given by the formula:
Figure RE-GDA0003577100850000037
the invention has the following beneficial effects: the invention provides an uncertainty-based monocular depth estimation method, which solves the problem that the detail depth around a depth map object is unclear, and the most advanced performance can be obtained on a KITTI data set.
In order to make the aforementioned and other objects, features and advantages of the invention comprehensible, preferred embodiments accompanied with figures are described in detail below.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts.
FIG. 1 is a basic diagram of the Retinex algorithm of the present invention;
FIG. 2 is a schematic diagram of an unsupervised monocular depth estimation method of the present invention;
FIG. 3 is a schematic diagram of the DispNet network structure of the present invention;
FIG. 4 is a graph I of the experimental results of the present invention;
FIG. 5 is a second graph showing the results of the experiment according to the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
An unsupervised monocular depth estimation method based on uncertain analysis comprises the following steps:
step 1: an unsupervised depth estimation network based on uncertainty is provided, and the depth prediction precision in monocular depth estimation is improved;
step 2: based on the step 1, the confidence coefficient of the estimated depth is predicted through modeling uncertainty, meanwhile, the model prediction precision is improved, and the uncertainty of an output result is quantified;
and step 3: based on the step 1-2, a Retinex illumination theory is used for constructing a brightness loss function, and the interference of dynamic objects in the scene is solved through operation and conversion according to the basic theory of a Retinex algorithm.
Further, the step 1 specifically comprises: the uncertainty in neural networks is largely divided into two areas: model uncertainty and random uncertainty; the model uncertainty mainly refers to the uncertainty of model parameters, and when a plurality of models have good effects, the final model parameters need to be selected from the models; when the amount of input data is large enough, the model uncertainty becomes small; in general, training data is enough during training, so random uncertainty in uncertainty analysis accounts for a main part;
first, a likelihood function is defined:
p(y|y*)=N(y*2);
wherein y denotes the depth of observation, y*Representing the depth, σ, of the model output2Representing an observed noise variance;
secondly, solving the likelihood function:
Figure RE-GDA0003577100850000051
then, an objective function is established:
Figure RE-GDA0003577100850000052
depth estimation of images is a regression task, L1Is superior to L2A penalty function; because L is1The optimization effect on small prediction errors is good, and the optimization just accords with the characteristic of depth estimation; therefore, the loss function for the uncertainty analysis is as follows:
Figure RE-GDA0003577100850000053
finally, the resulting loss function is:
L=LR+LS+LU
further, referring to FIG. 2, two consecutive frames I sampled from a given unmarked videotAnd It-1First, estimate its depth map D using a depth networktAnd Dt-1Then using the pose network PabTraining relative 6D poses between cameras; depth map D with predictiontAnd relative camera pose PabThe invention uses differentiable bilinear interpolation to convert It-1Is transformed to synthesize
Figure RE-GDA0003577100850000054
Similarly, an image is obtained
Figure RE-GDA0003577100850000055
Finally, will
Figure RE-GDA0003577100850000056
Input into a depth net to obtain
Figure RE-GDA0003577100850000057
By means of uncertainty analysis, in
Figure RE-GDA0003577100850000058
And
Figure RE-GDA0003577100850000059
form a loss function L therebetweenU
Further, for the depth estimation network, the present invention improves on DispNet, which takes a single RGB image as input and outputs a depth map.
Further, referring to fig. 3, for the posture network, the present invention uses a network DispNet without mask predicted branches.
Further, the step 3 specifically includes: the algorithm of Retinex is based on three assumptions:
(1) the real physical world is colorless and the colors seen by humans are the result of the interaction of natural light with matter in the objective world.
(2) Each color range in the object is composed of red, green, and blue of a certain wavelength.
(3) The color of each region is determined by red, green and blue.
Referring to fig. 1, a Retinex algorithm is different from a linear and nonlinear method that only can enhance a certain feature of an image, Retinex can achieve balance in three aspects of compression, edge enhancement and color normalization in a dynamic range, so that different types of images can be adaptively enhanced, and color fidelity, image edge portion enhancement and dynamic range compression can be obviously improved.
Through analyzing the illumination component and the reflection component, people can judge the illumination information and the reflection information in the image, thereby solving the problem of image brightness and color change caused by illumination change, particularly, the illumination information influencing the vision of people is removed through various conversion methods, the reflection information of an object is retained to the maximum extent, and because the attribute information of the object is contained in the reflection component, a human visual system imitating the human visual system develops a Retinex algorithm all the time, the Retinex algorithm is improved from a single-scale Retinex algorithm to a multiscale weighted average Retinex algorithm, and then the Retinex algorithm is developed into a color recovery multiscale Retinex algorithm;
firstly, according to the basic theory of Retinex algorithm, the following expression is obtained:
I(x,y)=R(x,y)×L(x,y);
secondly, the main principle of the single-scale Retinex algorithm is to perform convolution operation on three channels of the image and a central surrounding function respectively, and the image after the convolution operation is regarded as the estimation of the illumination component of the original image;
the process of solving for the incident component by convolution operation with a low pass filter can be expressed as:
L(x,y)=I(x,y)*G(x,y);
thirdly, the mathematical solving process of the single-scale Retinex algorithm is shown as follows:
Figure RE-GDA0003577100850000061
wherein i belongs to R, G and B and represents R, G and B color channels, Ri(x, y) represents the pixel value of the reflected image in the ith color channel, Ii(x, y) represents the pixel value of the original image I at the ith color channel (x, y), which represents a gaussian convolution operation, and G (x, y) represents a gaussian surround function;
Figure RE-GDA0003577100850000062
where σ represents the standard deviation in a gaussian function, referred to herein as the scale function; the size of the standard deviation has great influence on the Retinex algorithm; the smaller the sigma is, the better detail information of the enhanced image can be obtained, but image distortion and halo phenomena are easy to occur; the larger the sigma is, the better the color of the enhanced image is kept, but the image has a larger sharpening degree and a poorer contrast enhancement effect;
then, the depth estimation is a regression task, and the most commonly used loss functions in the optimization of the regression task comprise an L2 loss function and an L1 loss function;
Figure RE-GDA0003577100850000063
Figure RE-GDA0003577100850000064
the square operation makes the L2 loss function sensitive to abnormal values, has good optimization effect on large prediction errors, and has poor capability of further optimization on small prediction errors; the L1 loss function has a good optimization effect on small prediction errors, the optimization effect on large prediction errors is general, and the effect of the L1 loss function is slightly excellent in actual training; the uncertainty Loss function provided by the invention combines L1 Loss and heteroscedastic random uncertainty in a neural network, and consists of a residual regression term and an uncertainty regularization term;
finally, from the above, the conversion formula yields:
Figure RE-GDA0003577100850000071
Figure RE-GDA0003577100850000072
where N represents a pixel in the image, ri(x, y) represents incident light of an image.
Further, because luminosity loss does not provide sufficient information in low-texture or uniform regions of a scene, existing work incorporates smoothness prior to regularizing the estimated depth map; the edge-perceived smoothness loss used is given by:
Figure RE-GDA0003577100850000073
the invention carries out the experiment of monocular depth estimation, which comprises the following steps: the experimental environment provided by the invention comprises a software environment and a hardware environment, wherein the software environment comprises: windows 1064-bit operating system, CUDA 9.1, cuDNN 7.1, Pythrch deep learning framework, Python 3.7.0 and MATLAB R2018 a; the hardware environment is as follows: intel (R) core (TM) i7-7700 CPU @3.60GHz processor, 32GB RAM and NVIDIA GeForce GTX 1080Ti GPU, 11 GB;
according to the invention, a KITTI data set is adopted to carry out monocular depth estimation experiments, about 42000 pictures are included in a training set, and no ground true value exists; in the training process, a random gradient descent method is adopted for optimization solution, the weight parameters of the training network are continuously updated by using a back propagation algorithm, the initial learning rate is set to be 0.001, the momentum factor is set to be 0.9, and the weight attenuation factor is set to be 0.0005; the learning rate is closely related to the convergence speed of the training network, the network model cannot be converged if the learning rate is too large, and the convergence speed of the network model becomes slow if the learning rate is too small; in the invention, the maximum iteration number of the training network is 20000, the learning rate of the previous 12000 times is set to be 0.001, the learning rate of the previous 12000 times to 16000 times is set to be 0.0001, and the learning rate of the previous 16000 times is set to be 0.00001;
in order to objectively evaluate the proposed monocular depth estimation model, the model is quantitatively analyzed using the following four criteria:
mean absolute relative error (Rel):
Figure RE-GDA0003577100850000074
root mean square error: (RMS):
Figure RE-GDA0003577100850000075
mean log10 error (RMSlog):
accuracy within the threshold range:
Figure RE-GDA0003577100850000076
referring to fig. 4, through the unsupervised depth estimation method provided by the invention, experiments are carried out on random pictures in a test set, the topmost picture in fig. 4 is an original picture, the bottommost picture is the algorithm provided by the invention, and at the position of a dotted line frame in fig. 4, the phenomenon that the depth of the algorithm provided by the invention does not excessively drift around traffic lights is obviously seen, the precision is improved, and the reason of dynamic object interference is also solved;
referring to fig. 5, four pictures of adjacent frames are also realized by the algorithm provided by the present invention, and it can be easily seen that the depth around the traffic light is well improved.
Figure RE-GDA0003577100850000081
Referring to tables 1 and 2, to further verify the utility of the present invention, the following are respectively: three methods of basic network, basic network and brightness loss, basic network and uncertainty, basic network and brightness loss and uncertainty are used for carrying out experiments, and the three experimental methods have the same smoothness loss; the table data shows that the problem of low precision is solved by the brightness loss and the uncertainty loss in the invention;
TABLE 1 ablation experiment 1 (resolution of input image: 416X128)
Figure RE-GDA0003577100850000082
TABLE 2 ablation experiment 2 (resolution of input image: 832x256)
Figure RE-GDA0003577100850000083
The principle and the implementation mode of the invention are explained by applying specific embodiments in the invention, and the description of the embodiments is only used for helping to understand the method and the core idea of the invention; meanwhile, for a person skilled in the art, according to the idea of the present invention, there may be variations in the specific embodiments and the application scope, and in summary, the content of the present specification should not be construed as a limitation to the present invention.

Claims (7)

1. An unsupervised monocular depth estimation method based on uncertain analysis is characterized by comprising the following steps of:
step 1: an unsupervised depth estimation network based on uncertainty is provided, and the depth prediction precision in monocular depth estimation is improved;
step 2: based on the step 1, the confidence coefficient of the estimated depth is predicted through modeling uncertainty, meanwhile, the model prediction precision is improved, and the uncertainty of an output result is quantized;
and step 3: based on the step 1-2, a Retinex illumination theory is used for constructing a brightness loss function, and the interference of dynamic objects in the scene is solved through operation and conversion according to the basic theory of a Retinex algorithm.
2. The unsupervised monocular depth estimation method based on uncertainty analysis of claim 1, further comprising:
the step 1 specifically comprises the following steps:
first, a likelihood function is defined:
Figure DEST_PATH_IMAGE002
wherein
Figure DEST_PATH_IMAGE004
Which represents the depth of the observation(s),
Figure DEST_PATH_IMAGE006
the depth of the output of the model is represented,
Figure DEST_PATH_IMAGE008
representing an observed noise variance;
secondly, solving the likelihood function:
Figure 1
again, an objective function is established:
Figure DEST_PATH_IMAGE012
the depth estimation of the image is a regression task,
Figure DEST_PATH_IMAGE014
is superior to
Figure DEST_PATH_IMAGE016
A penalty function;
then, the loss function for the uncertainty analysis is as follows:
Figure DEST_PATH_IMAGE018
finally, the resulting loss function is:
Figure DEST_PATH_IMAGE020
3. the unsupervised monocular depth estimation method based on uncertainty analysis of claim 2, further comprising: two consecutive frames sampled from a given unmarked video
Figure DEST_PATH_IMAGE022
And
Figure DEST_PATH_IMAGE024
first, estimate its depth map using a depth network
Figure DEST_PATH_IMAGE026
And
Figure DEST_PATH_IMAGE028
then using the attitude network
Figure DEST_PATH_IMAGE030
Training relative 6D poses between the cameras; depth map with prediction
Figure 807568DEST_PATH_IMAGE026
And relative camera pose
Figure 193550DEST_PATH_IMAGE030
By differentiable bilinear interpolation
Figure 442128DEST_PATH_IMAGE024
Is formed by deformation
Figure DEST_PATH_IMAGE032
Similarly, an image is obtained
Figure DEST_PATH_IMAGE034
Finally, will (
Figure DEST_PATH_IMAGE036
) Input into a depth net to obtain
Figure 2
By means of uncertainty analysis, in
Figure DEST_PATH_IMAGE040
And
Figure DEST_PATH_IMAGE042
form a loss function between
Figure DEST_PATH_IMAGE044
4. The unsupervised monocular depth estimation method based on uncertainty analysis of claim 1, further comprising: for the depth estimation network, the present invention improves on DispNet, which takes a single RGB image as input and outputs a depth map.
5. The unsupervised monocular depth estimation method based on uncertainty analysis of claim 3, further comprising: for the posture network, the invention uses the network DispNet without mask predicted branches.
6. The unsupervised monocular depth estimation method based on uncertainty analysis of claim 1, further comprising: the step 3 specifically comprises the following steps:
firstly, according to the basic theory of Retinex algorithm, the following expression is obtained:
Figure DEST_PATH_IMAGE046
secondly, the process of solving the incident component by convolution operation with a low-pass filter can be expressed as:
Figure DEST_PATH_IMAGE048
thirdly, the mathematical solving process of the single-scale Retinex algorithm is shown as follows:
Figure DEST_PATH_IMAGE050
wherein the content of the first and second substances,
Figure DEST_PATH_IMAGE052
indicating the color channels of R, G, B,
Figure DEST_PATH_IMAGE054
indicating a reflected image is on
Figure DEST_PATH_IMAGE056
The pixel values of the individual color channels,
Figure DEST_PATH_IMAGE058
representing an original image
Figure DEST_PATH_IMAGE060
In the first place
Figure 729016DEST_PATH_IMAGE056
Color channel
Figure DEST_PATH_IMAGE062
The pixel values at, represent a gaussian convolution operation,
Figure DEST_PATH_IMAGE064
representing a gaussian surround function;
Figure DEST_PATH_IMAGE066
wherein the content of the first and second substances,
Figure DEST_PATH_IMAGE068
represents the standard deviation in a gaussian function, referred to herein as a scale function;
then, the depth estimation is a regression task, and the most commonly used loss functions in the optimization of the regression task comprise an L2 loss function and an L1 loss function;
Figure DEST_PATH_IMAGE070
Figure DEST_PATH_IMAGE072
finally, from the above, the conversion formula yields:
Figure DEST_PATH_IMAGE074
Figure DEST_PATH_IMAGE076
wherein N represents a pixel in the image,
Figure DEST_PATH_IMAGE078
the incident light representing an image.
7. The unsupervised monocular depth estimation method based on uncertainty analysis of claim 1, further comprising:
smoothness is included before regularizing the estimated depth map, and the used edge-perceived smoothness loss is expressed by the following formula:
Figure DEST_PATH_IMAGE080
CN202111185472.5A 2021-10-12 2021-10-12 Unsupervised monocular depth estimation method based on uncertain analysis Pending CN114549297A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111185472.5A CN114549297A (en) 2021-10-12 2021-10-12 Unsupervised monocular depth estimation method based on uncertain analysis

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111185472.5A CN114549297A (en) 2021-10-12 2021-10-12 Unsupervised monocular depth estimation method based on uncertain analysis

Publications (1)

Publication Number Publication Date
CN114549297A true CN114549297A (en) 2022-05-27

Family

ID=81668506

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111185472.5A Pending CN114549297A (en) 2021-10-12 2021-10-12 Unsupervised monocular depth estimation method based on uncertain analysis

Country Status (1)

Country Link
CN (1) CN114549297A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114782782A (en) * 2022-06-20 2022-07-22 武汉大学 Uncertainty quantification method for learning performance of monocular depth estimation model
CN114820755A (en) * 2022-06-24 2022-07-29 武汉图科智能科技有限公司 Depth map estimation method and system

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114782782A (en) * 2022-06-20 2022-07-22 武汉大学 Uncertainty quantification method for learning performance of monocular depth estimation model
CN114782782B (en) * 2022-06-20 2022-10-04 武汉大学 Uncertainty quantification method for learning performance of monocular depth estimation model
CN114820755A (en) * 2022-06-24 2022-07-29 武汉图科智能科技有限公司 Depth map estimation method and system

Similar Documents

Publication Publication Date Title
Li et al. Image dehazing using residual-based deep CNN
CN108986050B (en) Image and video enhancement method based on multi-branch convolutional neural network
CN109685072B (en) Composite degraded image high-quality reconstruction method based on generation countermeasure network
Negru et al. Exponential contrast restoration in fog conditions for driving assistance
CN103914813B (en) The restored method of colored haze image defogging and illumination compensation
KR102261532B1 (en) Method and system for image dehazing using single scale image fusion
CN110544213A (en) Image defogging method based on global and local feature fusion
CN109389569B (en) Monitoring video real-time defogging method based on improved DehazeNet
CN114549297A (en) Unsupervised monocular depth estimation method based on uncertain analysis
CN115063329A (en) Visible light and infrared image fusion enhancement method and system under low-illumination environment
CN116797488A (en) Low-illumination image enhancement method based on feature fusion and attention embedding
Agrawal et al. A comprehensive review on analysis and implementation of recent image dehazing methods
CN111476213A (en) Method and device for filling covering area of shelter based on road image
CN112703532B (en) Image processing method, device, equipment and storage medium
CN110969171A (en) Image classification model, method and application based on improved convolutional neural network
CN113284061B (en) Underwater image enhancement method based on gradient network
CN114170286B (en) Monocular depth estimation method based on unsupervised deep learning
CN115775376A (en) Crowd counting method based on low-light image enhancement
CN106657948A (en) low illumination level Bayer image enhancing method and enhancing device
CN116596792B (en) Inland river foggy scene recovery method, system and equipment for intelligent ship
Moghimi et al. Real-time underwater image resolution enhancement using super-resolution with deep convolutional neural networks
CN110223240A (en) Image defogging method, system and storage medium based on color decaying priori
CN115131229A (en) Image noise reduction and filtering data processing method and device and computer equipment
Liang et al. Learning to remove sandstorm for image enhancement
CN109064444B (en) Track slab disease detection method based on significance analysis

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination