CN112990299A - Depth map acquisition method based on multi-scale features, electronic device and storage medium - Google Patents

Depth map acquisition method based on multi-scale features, electronic device and storage medium Download PDF

Info

Publication number
CN112990299A
CN112990299A CN202110265024.XA CN202110265024A CN112990299A CN 112990299 A CN112990299 A CN 112990299A CN 202110265024 A CN202110265024 A CN 202110265024A CN 112990299 A CN112990299 A CN 112990299A
Authority
CN
China
Prior art keywords
scale
feature
map
feature map
block
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110265024.XA
Other languages
Chinese (zh)
Other versions
CN112990299B (en
Inventor
常青玲
崔岩
杨鑫
任飞
戴成林
胡昊
李敏华
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuyi University
Original Assignee
Wuyi University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuyi University filed Critical Wuyi University
Priority to CN202110265024.XA priority Critical patent/CN112990299B/en
Publication of CN112990299A publication Critical patent/CN112990299A/en
Application granted granted Critical
Publication of CN112990299B publication Critical patent/CN112990299B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computational Linguistics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

The invention provides a depth map acquisition method, a device and a storage medium based on multi-scale features, wherein the method comprises the following steps: acquiring an input image, and acquiring a multi-scale feature block according to the input image; obtaining a pooling feature map according to the multi-scale feature block, and obtaining a channel attention block according to the pooling feature map, wherein the channel attention block represents the relation among a plurality of features; obtaining an original fusion feature map according to the multi-scale feature block, and obtaining a target fusion feature map according to the original fusion feature map and the channel attention block; and splicing the target fusion characteristic graph and the original special effect graph and carrying out decoding operation to obtain a target depth graph. According to the scheme provided by the embodiment of the invention, the relation among the multi-scale features can be enhanced through the channel attention block, so that the multi-scale features can express the object information better, and the definition of the object information of the target depth map is effectively improved.

Description

Depth map acquisition method based on multi-scale features, electronic device and storage medium
Technical Field
The present invention relates to, but not limited to, the field of image processing, and in particular, to a method, an apparatus, and a storage medium for obtaining a depth map based on multi-scale features.
Background
With the development of image processing technology, the application of depth maps is more and more extensive. The depth map can be obtained by processing based on an image captured by a terminal, and generally requires input images at a plurality of angles in order to improve the definition of the depth map and the accuracy of the object structure. However, the multi-angle input image requires a terminal to be equipped with a plurality of cameras, and the cost of hardware is high. For a monocular camera, although some depth learning algorithms can obtain a depth map based on an input image of the monocular camera, since only one camera is provided, the angle of the obtained input image is limited, and a conventional image processing method usually loses a large amount of object information, so that the depth map is blurred, and the object structure is unclear.
Disclosure of Invention
The following is a summary of the subject matter described in detail herein. This summary is not intended to limit the scope of the claims.
The embodiment of the invention provides a depth map acquisition method, depth map acquisition equipment and a storage medium based on multi-scale features, which can improve the definition of object information of a depth map.
In a first aspect, an embodiment of the present invention provides a depth map obtaining method based on multi-scale features, including:
acquiring an input image, and acquiring a multi-scale feature block according to the input image;
obtaining a pooling feature map according to the multi-scale feature block, and obtaining a channel attention block according to the pooling feature map, wherein the channel attention block represents the relation among a plurality of features;
obtaining an original fusion feature map according to the multi-scale feature block, and obtaining a target fusion feature map according to the original fusion feature map and the channel attention block;
and decoding the target fusion feature map to obtain a target depth map.
The embodiment of the invention comprises the following steps: acquiring an input image, and acquiring a multi-scale feature block according to the input image; obtaining a pooling feature map according to the multi-scale feature block, and obtaining a channel attention block according to the pooling feature map, wherein the channel attention block represents the relation among a plurality of features; obtaining an original fusion feature map according to the multi-scale feature block, and obtaining a target fusion feature map according to the original fusion feature map and the channel attention block; and decoding the target fusion feature map to obtain a target depth map. According to the scheme provided by the embodiment of the invention, the relation among the multi-scale features can be enhanced through the channel attention block, so that the multi-scale features can express the object information better, and the definition of the object information of the target depth map is effectively improved.
As a further improvement of the present invention, the obtaining a multi-scale feature block according to the input image includes:
obtaining initial characteristics according to the input image;
acquiring a preset multi-scale feature fusion network, and performing feature aggregation on the initial features through the multi-scale feature fusion network to obtain a plurality of aggregation features of different scales;
stitching a plurality of the aggregated features to obtain the multi-scale feature block.
As a further improvement of the present invention, before the stitching a plurality of the aggregated features to obtain the multi-scale feature block, the method further comprises:
compressing a plurality of the aggregated features to the same number of channels.
As a further improvement of the present invention, the obtaining a pooling feature map according to the multi-scale feature block and obtaining a channel attention block according to the pooling feature map includes:
performing global pooling on the multi-scale feature blocks to obtain a pooled feature map;
and sequentially performing a compression operation and an activation operation on the pooled feature map to obtain the channel attention block.
As a further improvement of the present invention, before obtaining the target fused feature map according to the original fused feature map and the channel attention block, the method further comprises:
compressing the original fused feature map such that the number of channels of the original fused feature map is the same as the number of channels of the pooled feature map.
As a further improvement of the present invention, the decoding operation performed on the target fusion feature map to obtain a target depth map includes:
performing channel connection according to the target fusion feature map and the original fusion feature map to obtain a reference scale feature block;
compressing the reference scale feature block at least twice to obtain a plurality of initial depth maps with different scales;
and summing pixels of the plurality of initial depth maps to obtain the target depth map.
As a further improvement of the present invention, the pixel summing a plurality of the initial depth maps comprises:
acquiring a preset self-adaptive weight, wherein the self-adaptive weight corresponds to each scale;
performing pixel summation according to the adaptive weight and the plurality of initial depth maps, wherein the formula of the pixel summation is as follows: d ═ Σk∈lwk × dk; and D is the target depth map, wherein dk is the initial depth map of the kth scale, wk is the adaptive weight of the kth scale, and D is the target depth map.
As a further improvement of the invention, the sum of all the adaptive weights is 1.
In a second aspect, an embodiment of the present invention further provides an apparatus, including a memory, a processor, and a computer program stored on the memory and executable on the processor, where the processor, when executing the computer program, implements the method for acquiring a depth map based on multi-scale features according to the first aspect.
Additional features and advantages of the invention will be set forth in the description which follows, and in part will be obvious from the description, or may be learned by practice of the invention. The objectives and other advantages of the invention will be realized and attained by the structure particularly pointed out in the written description and claims hereof as well as the appended drawings.
Drawings
The accompanying drawings are included to provide a further understanding of the invention and are incorporated in and constitute a part of this specification, illustrate embodiments of the invention and together with the example serve to explain the principles of the invention and not to limit the invention.
Fig. 1 is a flowchart of a depth map acquisition method based on multi-scale features according to an embodiment of the present invention;
FIG. 2 is a flow chart for obtaining a multi-scale feature block according to another embodiment of the present invention;
FIG. 3 is a block diagram of a multi-scale fusion feature network according to another embodiment of the present invention;
FIG. 4 is a block diagram of a channel attention block provided in accordance with another embodiment of the present invention;
FIG. 5 is a flow diagram of a compression aggregation feature provided by another embodiment of the present invention;
FIG. 6 is a flow chart of a get channel attention block provided by another embodiment of the present invention;
FIG. 7 is a flow diagram of compressing a raw fused feature map as provided by another embodiment of the present invention;
FIG. 8 is a flow chart for obtaining a target depth map according to another embodiment of the present invention;
fig. 9 is a diagram of a network architecture for decoding provided by another embodiment of the present invention;
FIG. 10 is a flow chart of pixel summing provided by another embodiment of the present invention;
FIG. 11 is a diagram of an overall network architecture provided by another embodiment of the present invention;
fig. 12 is a device diagram of an apparatus provided by another embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is described in further detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
It should be noted that although functional blocks are partitioned in a schematic diagram of an apparatus and a logical order is shown in a flowchart, in some cases, the steps shown or described may be performed in a different order than the partitioning of blocks in the apparatus or the order in the flowchart. The terms "first," "second," and the like in the description, in the claims, or in the drawings described above, are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order.
The embodiments of the present invention will be further explained with reference to the drawings.
As shown in fig. 1, fig. 1 is a depth map obtaining method based on multi-scale features according to an embodiment of the present invention, including:
step S110, acquiring an input image, and obtaining a multi-scale fusion feature block according to the input image;
step S120, a pooling feature map is obtained according to the multi-scale feature block, and a channel attention block is obtained according to the pooling feature map, wherein the channel attention block represents the relation among a plurality of features;
step S130, obtaining an original fusion feature map according to the multi-scale feature block, and obtaining a target fusion feature map according to the original fusion feature map and the channel attention block;
and step S140, decoding the target fusion feature map to obtain a target depth map.
It should be noted that the input image may be a photograph or an image containing information of any object, and may be acquired by a monocular camera, and the specific content in the image is not limited in this embodiment. It can be understood that, a method in the prior art may be adopted to obtain a multi-scale feature block according to an input image, and this embodiment does not limit a specific multi-scale feature obtaining method, and may obtain a feature block composed of features of multiple scales.
It should be noted that, by obtaining the pooled feature map through pooling, features can be compressed, complexity is reduced, and features other than object information are reduced, so that the obtained channel attention block represents the relationship among a plurality of features, thereby achieving the effect of strengthening the relationship among the features during fusion. For example, a Dense Feature Fusion Network (DFFN) shown in fig. 3 may be adopted, and the structure shown in fig. 3 is only an example and does not limit the technical solution of the present application. Can be used for
It can be understood that for the image acquired by the monocular camera, the relevance of the object information in the target depth map obtained by decoding is stronger by strengthening the relation between the features, the definition of the depth map is effectively improved, and more accurate object information is obtained.
It is to be understood that the target fused feature map may be obtained by performing a dot product on the original fused feature map and the channel attention block, or may be obtained by performing other methods that can achieve the same effect, which is not limited herein.
It should be noted that any type of network may be used for decoding the target Fusion feature map, and a multi-scale Depth map Fusion Module (DAFM) is preferably used in this embodiment, and a structure of the dapm is shown in fig. 9, where it should be noted that the structure shown in fig. 9 is merely an example, and does not limit the technical solution of the present application, and the multi-scale Depth map Fusion Module can simultaneously decode and adaptively fuse features of multiple scales, and can effectively improve decoding efficiency in the case of many parameters.
In addition, referring to fig. 2, in an embodiment, step S110 in the embodiment shown in fig. 2 further includes, but is not limited to, the following steps:
step S210, obtaining initial characteristics according to an input image;
step S220, acquiring a preset multi-scale feature fusion network, and performing feature aggregation on the initial features through the multi-scale feature fusion network to obtain a plurality of aggregation features with different scales;
and step S230, splicing the multiple aggregation features to obtain a multi-scale feature block.
It should be noted that, on the premise of the DFFN shown in fig. 3, 5 approximate sub-modules are adopted to perform upsampling and downsampling on the initial features of the input image, and then perform feature aggregation on the initial features to obtain aggregated features, and through aggregation operation, information loss in the feature extraction process can be effectively reduced, so that the obtained target depth map can reflect more object information.
It is noted that, because the connection among a plurality of features gradually fades after a plurality of convolution upsampling and downsampling, the connection among the features can be further strengthened through the channel attention block, and the definition of the target depth map is ensured.
In addition, referring to fig. 5, in an embodiment, before performing step S230 in the embodiment shown in fig. 2, the following steps are further included, but not limited to:
step S510, compress the multiple aggregated features to the same number of channels.
It should be noted that, since the aggregation features need to be spliced, in order to ensure the accuracy of splicing, the aggregation features need to be compressed before splicing to ensure that the number of channels is the same.
In addition, referring to fig. 6, in an embodiment, step S120 in the embodiment shown in fig. 1 further includes, but is not limited to, the following steps:
step S610, performing global pooling on the multi-scale feature blocks to obtain a pooled feature map;
and step S620, sequentially performing compression operation and activation operation on the pooled feature map to obtain a channel attention block.
It should be noted that after obtaining the multi-scale feature block from the concatenation in the DFFN, the compressing operation and the activating operation can strengthen the features with value and eliminate the useless information, and the specific compressing operation and the activating operation can refer to the following examples:
referring to fig. 4, using a global average pooling layer to pool the multi-scale feature blocks into a 1 × 1 × C pooled feature map, compressing the pooled feature map by a 1 × 1 convolution, and then activating using a ReLU function, completing the first compression and activation; the pooled feature map is convolved again with a 1 × 1 convolution and activated using Sigmoid function without compression, resulting in the channel attention block shown in fig. 4.
In addition, referring to fig. 7, in an embodiment, step S130 in the embodiment shown in fig. 1 further includes, but is not limited to, the following steps:
step S710, compress the original fused feature map to make the number of channels of the original fused feature map the same as the number of channels of the pooled feature map.
It should be noted that the channel attention block is obtained from the pooled feature map, so the number of channels is the same as the pooled feature map, and the original fused feature map and the channel attention block need to be point-multiplied to obtain the target fused feature map, so the number of channels in the original fused feature map needs to be ensured to be the same as the number of channels in the pooled feature map by compression.
In addition, referring to fig. 8, in an embodiment, step S140 in the embodiment shown in fig. 1 further includes, but is not limited to, the following steps:
step S810, channel connection is carried out according to the target fusion feature map and the original fusion feature map to obtain a reference scale feature block;
step S820, compressing the reference scale feature block at least twice to obtain a plurality of initial depth maps with different scales;
step S830, performing pixel summation on the multiple initial depth maps to obtain a target depth map.
It should be noted that, as shown in fig. 11, the network structure diagram of this embodiment may obtain a target fusion feature diagram through DFFN after obtaining an original fusion feature diagram according to decomposition of an input image, and input the target fusion feature diagram and the target fusion feature diagram to DAFM for channel connection after channel connection to obtain a reference scale feature block.
It can be understood that, in order to improve the expression capability of the network, a step-by-step compression method may be used, and two convolutions of 3 × 3 are used to compress the reference scale feature block, so as to obtain the initial depth maps of multiple scales.
In addition, referring to fig. 10, in an embodiment, step S830 in the embodiment shown in fig. 8 further includes, but is not limited to, the following steps:
step S1010, acquiring preset self-adaptive weight, wherein the self-adaptive weight corresponds to each scale;
step S1020, performing pixel summation according to the adaptive weight and the plurality of initial depth maps, where a formula of the pixel summation is: d ═ Σk∈lwk × dk; and D is the target depth map.
It should be noted that an adaptive weight may be set for the initial depth map through a convolution of 1 × 1, and the adaptive weight may be learned through an existing standard back propagation mechanism, and may embody a characteristic weight, which is not limited in this embodiment. And finally, summing all the initial depth maps at the pixel level to obtain a final target depth map.
In the target depth map, the depth value Y of the (i, j) th pixeli,jSatisfies the following formula:
Figure BDA0002972106460000051
Figure BDA0002972106460000052
wherein
Figure BDA0002972106460000053
Is the depth value at (i, j) of the initial feature at the k-th scale. Alpha is alphai,j,βi,j,γi,ji,ji,jIs the adaptive weight corresponding to each scale.
In addition, in one embodiment, the sum of all the adaptive weights is 1.
It should be noted that, in order to avoid overfitting, the adaptive weights may be mathematically constrained such that the sum of all adaptive weights is 1For example, the adaptive weights in the above embodiments satisfy the following relation: alpha is alphai,ji,ji,ji,ji,j=1。
Additionally, referring to fig. 12, an embodiment of the present invention also provides an apparatus 1200 including: memory 1210, processor 1220, and computer programs stored on memory 1210 and operable on processor 1220.
The processor 1220 and the memory 1210 may be connected by a bus or other means.
Non-transitory software programs and instructions required to implement the multi-scale feature-based depth map acquisition method of the above-described embodiment are stored in the memory 1210, and when executed by the processor 1220, perform the multi-scale feature-based depth map acquisition method of the above-described embodiment, for example, the method steps S110 to S140 in fig. 1, the method steps S210 to S230 in fig. 2, the method step S510 in fig. 5, the method steps S610 to S620 in fig. 6, the method step S710 in fig. 7, the method steps S810 to S830 in fig. 8, and the method steps S1010 to S1020 in fig. 10, which are described above, are performed.
The above-described embodiments of the apparatus are merely illustrative, wherein the units illustrated as separate components may or may not be physically separate, i.e. may be located in one place, or may also be distributed over a plurality of network elements. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of the present embodiment.
Furthermore, an embodiment of the present invention further provides a computer-readable storage medium, where the computer-readable storage medium stores computer-executable instructions, which are executed by a processor or a controller, for example, by a processor in the above-mentioned apparatus embodiment, and enable the processor to perform the depth map acquisition method based on the multi-scale features in the above-mentioned embodiment, for example, the method steps S110 to S140 in fig. 1, the method steps S210 to S230 in fig. 2, the method step S510 in fig. 5, the method steps S610 to S620 in fig. 6, the method step S710 in fig. 7, the method steps S810 to S830 in fig. 8, and the method steps S1010 to S1020 in fig. 10, which are described above, are performed. One of ordinary skill in the art will appreciate that all or some of the steps, systems, and methods disclosed above may be implemented as software, firmware, hardware, and suitable combinations thereof. Some or all of the physical components may be implemented as software executed by a processor, such as a central processing unit, digital signal processor, or microprocessor, or as hardware, or as an integrated circuit, such as an application specific integrated circuit. Such software may be distributed on computer readable media, which may include computer storage media (or non-transitory media) and communication media (or transitory media). The term computer storage media includes volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data, as is well known to those of ordinary skill in the art. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, Digital Versatile Disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can accessed by a computer. In addition, communication media typically embodies computer readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media as known to those skilled in the art.
While the preferred embodiments of the present invention have been described in detail, it will be understood by those skilled in the art that the foregoing and various other changes, omissions and deviations in the form and detail thereof may be made without departing from the scope of this invention.

Claims (10)

1. A depth map acquisition method based on multi-scale features is characterized by comprising the following steps:
acquiring an input image, and acquiring a multi-scale feature block according to the input image;
obtaining a pooling feature map according to the multi-scale feature block, and obtaining a channel attention block according to the pooling feature map, wherein the channel attention block represents the relation among a plurality of features;
obtaining an original fusion feature map according to the multi-scale feature block, and obtaining a target fusion feature map according to the original fusion feature map and the channel attention block;
and decoding the target fusion feature map to obtain a target depth map.
2. The method of claim 1, wherein obtaining a multi-scale feature block from the input image comprises:
obtaining initial characteristics according to the input image;
acquiring a preset multi-scale feature fusion network, and performing feature aggregation on the initial features through the multi-scale feature fusion network to obtain a plurality of aggregation features of different scales;
stitching a plurality of the aggregated features to obtain the multi-scale feature block.
3. The method of claim 2, wherein prior to stitching the plurality of aggregated features to obtain the multi-scale feature block, the method further comprises:
compressing a plurality of the aggregated features to the same number of channels.
4. The method of claim 1, wherein obtaining a pooled feature map from the multi-scale feature block and a channel attention block from the pooled feature map comprises:
performing global pooling on the multi-scale feature blocks to obtain a pooled feature map;
and sequentially performing a compression operation and an activation operation on the pooled feature map to obtain the channel attention block.
5. The method of claim 1, wherein prior to said deriving a target fused feature map from said original fused feature map and said channel attention block, said method further comprises:
compressing the original fused feature map such that the number of channels of the original fused feature map is the same as the number of channels of the pooled feature map.
6. The method of claim 1, wherein the decoding the target fused feature map to obtain a target depth map comprises:
performing channel connection according to the target fusion feature map and the original fusion feature map to obtain a reference scale feature block;
compressing the reference scale feature block at least twice to obtain a plurality of initial depth maps with different scales;
and summing pixels of the plurality of initial depth maps to obtain the target depth map.
7. The method of claim 6, wherein said pixel summing a plurality of said initial depth maps comprises:
acquiring a preset self-adaptive weight, wherein the self-adaptive weight corresponds to each scale;
performing pixel summation according to the adaptive weight and the plurality of initial depth maps, wherein the formula of the pixel summation is as follows: d ═ Σk∈lwk × dk; and D is the target depth map, wherein dk is the initial depth map of the kth scale, wk is the adaptive weight of the kth scale, and D is the target depth map.
8. The method of claim 7, wherein: the sum of all the adaptive weights is 1.
9. An electronic device comprising a memory, a processor and a computer program stored on the memory and executable on the processor, wherein the processor implements the method for obtaining a depth map based on multi-scale features according to any one of claims 1 to 8 when executing the computer program.
10. A computer-readable storage medium storing computer-executable instructions for performing the multi-scale feature based depth map acquisition method according to any one of claims 1 to 8.
CN202110265024.XA 2021-03-11 2021-03-11 Depth map acquisition method based on multi-scale features, electronic equipment and storage medium Active CN112990299B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110265024.XA CN112990299B (en) 2021-03-11 2021-03-11 Depth map acquisition method based on multi-scale features, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110265024.XA CN112990299B (en) 2021-03-11 2021-03-11 Depth map acquisition method based on multi-scale features, electronic equipment and storage medium

Publications (2)

Publication Number Publication Date
CN112990299A true CN112990299A (en) 2021-06-18
CN112990299B CN112990299B (en) 2023-10-17

Family

ID=76335020

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110265024.XA Active CN112990299B (en) 2021-03-11 2021-03-11 Depth map acquisition method based on multi-scale features, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN112990299B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114693970A (en) * 2022-03-28 2022-07-01 北京百度网讯科技有限公司 Object classification method, deep learning model training method, device and equipment
WO2024000728A1 (en) * 2022-06-28 2024-01-04 五邑大学 Monocular three-dimensional plane recovery method, device, and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110188685A (en) * 2019-05-30 2019-08-30 燕山大学 A kind of object count method and system based on the multiple dimensioned cascade network of double attentions
CN111738110A (en) * 2020-06-10 2020-10-02 杭州电子科技大学 Remote sensing image vehicle target detection method based on multi-scale attention mechanism
CN112287940A (en) * 2020-10-30 2021-01-29 西安工程大学 Semantic segmentation method of attention mechanism based on deep learning
CN112396645A (en) * 2020-11-06 2021-02-23 华中科技大学 Monocular image depth estimation method and system based on convolution residual learning

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110188685A (en) * 2019-05-30 2019-08-30 燕山大学 A kind of object count method and system based on the multiple dimensioned cascade network of double attentions
CN111738110A (en) * 2020-06-10 2020-10-02 杭州电子科技大学 Remote sensing image vehicle target detection method based on multi-scale attention mechanism
CN112287940A (en) * 2020-10-30 2021-01-29 西安工程大学 Semantic segmentation method of attention mechanism based on deep learning
CN112396645A (en) * 2020-11-06 2021-02-23 华中科技大学 Monocular image depth estimation method and system based on convolution residual learning

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114693970A (en) * 2022-03-28 2022-07-01 北京百度网讯科技有限公司 Object classification method, deep learning model training method, device and equipment
WO2024000728A1 (en) * 2022-06-28 2024-01-04 五邑大学 Monocular three-dimensional plane recovery method, device, and storage medium

Also Published As

Publication number Publication date
CN112990299B (en) 2023-10-17

Similar Documents

Publication Publication Date Title
US11222211B2 (en) Method and apparatus for segmenting video object, electronic device, and storage medium
US20190020871A1 (en) Visual quality preserving quantization parameter prediction with deep neural network
WO2020098422A1 (en) Encoded pattern processing method and device , storage medium and electronic device
CN108764039B (en) Neural network, building extraction method of remote sensing image, medium and computing equipment
CN112990299A (en) Depth map acquisition method based on multi-scale features, electronic device and storage medium
US20220392019A1 (en) Method for denoising videos and electronic device therefor
CN113781320A (en) Image processing method and device, terminal equipment and storage medium
CN109035257A (en) portrait dividing method, device and equipment
CN111353965B (en) Image restoration method, device, terminal and storage medium
CN115984666A (en) Cross-channel pyramid pooling method and system, convolutional neural network and processing method
WO2022126333A1 (en) Image filling method and apparatus, decoding method and apparatus, electronic device, and medium
CN111815529A (en) Low-quality image classification enhancement method based on model fusion and data enhancement
CN115209150B (en) Video coding parameter acquisition method and device and electronic equipment
CN113727050B (en) Video super-resolution processing method and device for mobile equipment and storage medium
WO2019236347A1 (en) Prediction for light-field coding and decoding
CN114584781A (en) Video compression method and device and computer readable storage medium
CN114612316A (en) Method and device for removing rain from nuclear prediction network image
CN111382753B (en) Light field semantic segmentation method, system, electronic terminal and storage medium
CN114078096A (en) Image deblurring method, device and equipment
CN114299105A (en) Image processing method, image processing device, computer equipment and storage medium
CN114638748A (en) Image processing method, image restoration method, computer device, and storage medium
CN112669240A (en) High-definition image restoration method and device, electronic equipment and storage medium
CN110475044A (en) Image transfer method and device, electronic equipment, computer readable storage medium
CN116962689A (en) Method and device for dividing depth map coding units in three-dimensional video frame
CN114782256B (en) Image reconstruction method and device, computer equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant