CN114491122A - Graph matching method for searching similar images - Google Patents

Graph matching method for searching similar images Download PDF

Info

Publication number
CN114491122A
CN114491122A CN202111634430.5A CN202111634430A CN114491122A CN 114491122 A CN114491122 A CN 114491122A CN 202111634430 A CN202111634430 A CN 202111634430A CN 114491122 A CN114491122 A CN 114491122A
Authority
CN
China
Prior art keywords
edge
point
graph
matrix
information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202111634430.5A
Other languages
Chinese (zh)
Other versions
CN114491122B (en
Inventor
杨益枘
林旭滨
何力
管贻生
张宏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong University of Technology
Original Assignee
Guangdong University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong University of Technology filed Critical Guangdong University of Technology
Priority to CN202111634430.5A priority Critical patent/CN114491122B/en
Publication of CN114491122A publication Critical patent/CN114491122A/en
Application granted granted Critical
Publication of CN114491122B publication Critical patent/CN114491122B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Molecular Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Library & Information Science (AREA)
  • Databases & Information Systems (AREA)
  • Image Processing (AREA)
  • Editing Of Facsimile Originals (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention discloses a graph matching method for similar image retrieval, which mainly comprises two stages of off-line data set construction and on-line deep learning training: the first stage comprises selecting a Pascal VOC data set as a training data set; a number of images with annotation points covering all kinds of the data set are selected as a training set. The second stage comprises the following steps: adopting a pre-trained VGG-16 neural network as a feature extractor; each image generates a topological structure of a bidirectional edge through a full-connection Delaunay triangulation technology; after the point feature embedding of the topological geometric information is completed, the feature description of the edge is carried out on the basis of a point-edge correlation matrix; according to the edge feature description vectors of the respective graphs, an edge-to-edge similarity matrix can be constructed; through the steps, the final point characteristic sum can be obtained, and then the similarity matrix of point-point matching is calculated. The scheme also has the advantages of high retrieval performance, high efficiency and easy implementation.

Description

Graph matching method for searching similar images
Technical Field
The invention relates to the technical field of image retrieval, in particular to a graph matching method for similar image retrieval.
Background
With the development of the internet, how to efficiently retrieve images meeting the requirements of users in a network environment is a core technical problem. Generally, image retrieval techniques are mainly divided into two branches: text-based and content-based retrieval. Text-based image retrieval typically queries for images in the form of keywords, or looks through images under specific categories in the form of a hierarchical directory. And the content-based image retrieval is to retrieve other images with similar characteristics from an image database according to semantic content and characteristics of the images.
The existing content-based image retrieval system firstly extracts feature information of image content, stores the feature information in a feature library, and then compares and sorts relevant features according to the features of a query image to obtain a retrieval result of the image. The image retrieval technology based on the content uses a computer to carry out mathematical description of a unified rule on the image, reduces the manpower consumption of manually labeling the image keywords, and thus improves the retrieval efficiency. With the improvement of computer performance and the development of deep learning, the computer can extract abundant characteristics such as object color, shape and structure from the image. However, similarity matching of the structured feature information is a problem with high computational complexity.
From the perspective of mathematical optimization, graph matching of structured information belongs to the NP-hard second order composition problem. Graph matching aims to find the corresponding relation between nodes and nodes among objects by utilizing graph structure information. On the other hand, the booming deep learning and graph convolution neural networks have great potential in the graph matching problem. With graph embedding techniques based on graph convolution neural networks, a second order combinatorial problem, which was previously difficult to solve accurately in polynomial time, translates into a first order problem that can be solved accurately in polynomial time. However, the existing depth map matching method based on the map embedding technology does not consider the similarity information of the second-order edge and the edge, and the information is introduced as the cross-map embedding information, so that the precision and the efficiency are improved. For this reason, further improvements and refinements are required in the prior art.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provide a graph matching method for the retrieval of similar images.
The purpose of the invention is realized by the following technical scheme:
a graph matching method for image retrieval of the same kind mainly comprises two stages of off-line data set construction and on-line deep learning training, and comprises the following specific steps:
stage one: and constructing a data set matched with the offline depth image.
Step S1: the Pascal VOC dataset was chosen as the training dataset.
Step S2: a number of images with annotation points covering all kinds of the data set are selected as a training set.
And a second stage: and training a depth map matching network on line.
Step S3: and adopting a pre-trained VGG-16 neural network as a feature extractor, wherein the parameters of the neural network are pre-trained on the ImageNet data set.
Step S4: and each image generates a topological structure of the bidirectional edge through a full-connected Delaunay triangulation technology.
Step S5: after the point feature embedding of the topological geometric information is completed, the feature description of the edge is carried out on the basis of the point-edge correlation matrix.
Step S6: according to the edge feature description vectors of the respective graphs, an edge-to-edge similarity matrix K can be constructede
Step S61: the point-edge pairing relation of graph matching can be constructed into a relevance graph model.
Step S62: according to the topological structure of the association graph, the edge similarity score and the point similarity can be associated to obtain the cross-graph conversion matrix.
Step S63: and (4) carrying out cross-map point embedding operation by taking a cross-map distribution matrix as prior information.
Step S7: through the steps, the final point characteristics can be obtained
Figure BDA0003441409920000021
And
Figure BDA0003441409920000022
a similarity matrix for the point-to-point matches is then calculated.
As a preferable embodiment of the present invention, the step S3 further includes the steps of: two images to be matched are obtained by a feature extractor
Figure BDA0003441409920000023
And
Figure BDA0003441409920000024
where d is the dimension of the feature vector, n1And n2Respectively counting the number of the characteristic points of the two images; f1And F2The outputs extracted from the relu4_2 and relu5_1 layers of the VGG-16 neural network are spliced.
As a preferable embodiment of the present invention, the step S4 further includes the steps of: the attribute of each edge is formed by two end point coordinates after normalization, and the connection information of the edge represents the topological structure information of each graph; then, inputting the point characteristic information and the edge attribute information into a neural network SplineCNN of the graph as input information; SplineCNN is used as a geometric topological information embedding technology, and MAX aggregation is adopted in structural information aggregation; finally, point characteristics embedded with respective geometric topological information are obtained
Figure BDA0003441409920000031
And
Figure BDA0003441409920000032
as a preferable embodiment of the present invention, the step S5 further includes the steps of: the point-edge correlation matrices of the two graphs are respectively
Figure BDA0003441409920000033
And
Figure BDA0003441409920000034
wherein e1And e2Respectively representing the number of edges of the two figures, when Gi,k=Hj,kWhen 1, it indicates that the edge k starts from node i to node j ends; characteristics of the edge
Figure BDA0003441409920000035
And
Figure BDA0003441409920000036
is defined as follows:
Figure BDA0003441409920000037
as a preferable embodiment of the present invention, the step S6 further includes the steps of: edge-to-edge recognition matrix Ke
Figure BDA0003441409920000038
Wherein the content of the first and second substances,
Figure BDA0003441409920000039
is a training parameter; keEach element of the matrix represents edge-to-edge matching information, and in order to expand the difference between the similarity values of the edges, i.e. emphasize the high similarity values and compress the low similarity values, the Ke matrix is normalized to obtain a normalized epsilon matrix:
ε=softmax(Ke) Formula (3)
Then, the normalized epsilon matrix is passed through a companion graphIs converted into a cross-edge conversion matrix
Figure BDA00034414099200000310
Figure BDA00034414099200000311
Conversion matrix based on cross-graph
Figure BDA00034414099200000312
We can get the feature embedding information across the graph; for node
Figure BDA00034414099200000313
Figure BDA00034414099200000314
Cross-map feature information mj→iIs calculated as follows:
Figure BDA00034414099200000315
finally, vector addition operation is carried out on the cross-map feature information and the point feature information:
Figure BDA00034414099200000316
similar operations are also performed for the feature points of the second graph.
As a preferable embodiment of the present invention, the step S7 further includes the steps of: the similarity matrix formula is as follows:
Figure BDA00034414099200000317
the linear solution to the graph matching problem relies on the Sinkhorn iterative algorithm, which is sequentially normalized along rows and columns to obtain soft scores according to the obtained score matrix SMatching matrix
Figure BDA0003441409920000041
Pij=Sinkhorn(exp(Sij) Equation (8).
As a preferable aspect of the present invention, the graph matching method further includes step S8: given truth allocation matrix
Figure BDA0003441409920000042
And a soft distribution matrix P, wherein the error can be obtained by constructing a cross entropy loss function:
Figure BDA0003441409920000043
as a preferable embodiment of the present invention, the step S1 further includes the steps of: the data set contains several different categories of images: airplanes, bicycles, birds, boats, bottles, buses, cars, cats, chairs, cattle, tables, dogs, horses, motorcycles, people, plants, sheep, sofas, trains, televisions; each image contains 6 to 23 annotated feature point image coordinates.
As a preferable embodiment of the present invention, the step S2 further includes the steps of: 1682 sheets were selected accordingly as the test set. For each image to be trained, extracting a bounding box containing all annotation feature points, adjusting the size of the image to be 256 multiplied by 256, and finally entering the deep learning network training.
The working process and principle of the invention are as follows: aiming at the problem of precision loss caused by neglecting second-order edge and edge similarity information in the existing depth map matching scheme based on the map embedding technology, the second-order edge and edge similarity information is introduced by using a depth map matching model based on the cross-map embedding technology, and the method is applied to image retrieval of similar objects, so that the matching performance is improved, the cost on memory consumption is also obviously reduced, and the performance and the efficiency of image retrieval are greatly improved finally.
Drawings
Fig. 1 is a schematic flow chart of a graph matching method for the same kind of image retrieval according to the present invention.
Fig. 2 is a schematic diagram of a graph matching method for searching similar images according to the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention clearer and clearer, the present invention is further described below with reference to the accompanying drawings and examples.
Example 1:
as shown in fig. 1 to fig. 2, the present embodiment discloses a graph matching method for retrieving similar images, which mainly includes two stages of offline data set construction and online deep learning training, and the specific steps are as follows:
stage one: and constructing a data set matched with the offline depth image.
Step S1: the Pascal VOC dataset was chosen as the training dataset.
Step S2: a number of images with annotation points covering all kinds of the data set are selected as a training set.
And a second stage: and training the depth map matching network on line.
Step S3: and adopting a pre-trained VGG-16 neural network as a feature extractor, wherein the parameters of the neural network are pre-trained on the ImageNet data set.
Step S4: and each image generates a topological structure of the bidirectional edge through a full-connected Delaunay triangulation technology.
Step S5: after the point feature embedding of the topological geometric information is completed, the feature description of the edge is carried out on the basis of the point-edge correlation matrix.
Step S6: according to the edge feature description vectors of the respective graphs, an edge-to-edge similarity matrix K can be constructede
Step S61: the point-edge pairing relation of graph matching can be constructed into a relevance graph model.
Step S62: according to the topological structure of the association graph, the edge similarity score and the point similarity can be associated to obtain the cross-graph conversion matrix.
Step S63: and (4) carrying out cross-map point embedding operation by taking a cross-map distribution matrix as prior information.
Step S7: through the steps, the final point characteristics can be obtained
Figure BDA0003441409920000051
And
Figure BDA0003441409920000052
a similarity matrix for the point-to-point matches is then calculated.
As a preferable embodiment of the present invention, the step S3 further includes the steps of: two images to be matched are obtained by a feature extractor
Figure BDA0003441409920000053
And
Figure BDA0003441409920000054
where d is the dimension of the feature vector, n1And n2Respectively counting the number of the characteristic points of the two images; f1And F2The outputs extracted from the relu4_2 and relu5_1 layers of the VGG-16 neural network are spliced.
As a preferable embodiment of the present invention, the step S4 further includes the steps of: the attribute of each edge is formed by two end point coordinates after normalization, and the connection information of the edge represents the topological structure information of each graph; then, inputting the point characteristic information and the edge attribute information into a neural network SplineCNN of the graph as input information; SplineCNN is used as a geometric topological information embedding technology, and MAX aggregation is adopted in structural information aggregation; finally, point characteristics embedded with respective geometric topological information are obtained
Figure BDA0003441409920000055
And
Figure BDA0003441409920000056
as a preferable embodiment of the present invention, the step S5 further includes the steps of: of two figuresThe point-edge correlation matrix is respectively
Figure BDA0003441409920000057
And
Figure BDA0003441409920000058
wherein e1And e2Respectively representing the number of edges of the two figures, when Gi,k=Hj,kWhen 1, it indicates that the edge k starts from node i to node j ends; characteristics of the edge
Figure BDA0003441409920000061
And
Figure BDA0003441409920000062
is defined as follows:
Figure BDA0003441409920000063
as a preferable embodiment of the present invention, the step S6 further includes the steps of: edge-to-edge recognition matrix Ke
Figure BDA0003441409920000064
Wherein the content of the first and second substances,
Figure BDA0003441409920000065
is a training parameter; keEach element of the matrix represents edge-to-edge matching information, and in order to expand the difference between the similarity values of the edges, i.e. emphasize the high similarity values and compress the low similarity values, the Ke matrix is normalized to obtain a normalized epsilon matrix:
ε=softmax(Ke) Formula (3)
Then, the normalized epsilon matrix is converted into a cross-edge conversion matrix through the structure of the accompanying graph
Figure BDA0003441409920000066
Figure BDA0003441409920000067
Conversion matrix based on cross-graph
Figure BDA0003441409920000068
We can get the feature embedding information across the graph; for node
Figure BDA0003441409920000069
Figure BDA00034414099200000610
Cross-map feature information mj→iIs calculated as follows:
Figure BDA00034414099200000611
finally, vector addition operation is carried out on the cross-map feature information and the point feature information:
Figure BDA00034414099200000612
similar operations are also performed for the feature points of the second graph.
As a preferable embodiment of the present invention, the step S7 further includes the steps of: the similarity matrix formula is as follows:
Figure BDA00034414099200000613
the linear solution of the graph matching problem depends on a Sinkhorn iterative algorithm, and the algorithm is sequentially normalized along rows and columns according to the obtained fractional matrix S to obtain a soft distribution matrix
Figure BDA00034414099200000614
Pij=Sinkhorn(exp(Sij) Equation (8).
As a preferable aspect of the present invention, the graph matching method further includes step S8: given truth-value distribution matrix
Figure BDA0003441409920000071
And a soft distribution matrix P, wherein the error can be obtained by constructing a cross entropy loss function:
Figure BDA0003441409920000072
as a preferable embodiment of the present invention, the step S1 further includes the steps of: the data set contains several different categories of images: airplanes, bicycles, birds, boats, bottles, buses, cars, cats, chairs, cattle, tables, dogs, horses, motorcycles, people, plants, sheep, sofas, trains, televisions; each image contains 6 to 23 annotated feature point image coordinates.
As a preferable embodiment of the present invention, the step S2 further includes the steps of: 1682 sheets were selected accordingly as the test set. For each image to be trained, extracting a bounding box containing all annotation feature points, adjusting the size of the image to be 256 multiplied by 256, and finally entering the deep learning network training.
The working process and principle of the invention are as follows: aiming at the problem of precision loss caused by neglecting second-order edge and edge similarity information in the existing depth map matching scheme based on the map embedding technology, the second-order edge and edge similarity information is introduced by using a depth map matching model based on the cross-map embedding technology, and the method is applied to image retrieval of similar objects, so that the matching performance is improved, the cost on memory consumption is also obviously reduced, and the performance and the efficiency of image retrieval are greatly improved finally.
The above embodiments are preferred embodiments of the present invention, but the present invention is not limited to the above embodiments, and any other changes, modifications, substitutions, combinations, and simplifications which do not depart from the spirit and principle of the present invention should be construed as equivalents thereof, and all such changes, modifications, substitutions, combinations, and simplifications are intended to be included in the scope of the present invention.

Claims (9)

1. A graph matching method for image retrieval of the same kind is characterized in that the graph matching method mainly comprises two stages of off-line data set construction and on-line deep learning training, and the specific steps are as follows:
stage one: constructing a data set matched with the offline depth image;
step S1: selecting a Pascal VOC data set as a training data set;
step S2: selecting a plurality of images which are provided with annotation points and cover all kinds of data sets as training sets;
and a second stage: training a depth map matching network on line;
step S3: adopting a pre-trained VGG-16 neural network as a feature extractor, wherein parameters of the neural network are pre-trained on an ImageNet data set;
step S4: each image generates a topological structure of a bidirectional edge through a full-connection Delaunay triangulation technology;
step S5: after the point feature embedding of the topological geometric information is completed, the feature description of the edge is carried out on the basis of a point-edge correlation matrix;
step S6: according to the edge feature description vectors of the respective graphs, an edge-to-edge similarity matrix K can be constructede
Step S61: point-edge pairing relation of graph matching can be constructed into an association graph model;
step S62: according to the topological structure of the association graph, the edge similarity score and the point similarity can be associated to obtain a cross-graph conversion matrix;
step S63: performing cross-map point embedding operation by taking a cross-map distribution matrix as prior information;
step S7: through the steps, the final point characteristics can be obtained
Figure FDA0003441409910000011
And
Figure FDA0003441409910000012
a similarity matrix for the point-to-point matches is then calculated.
2. The graph matching method for homogeneous image retrieval according to claim 1, wherein said step S3 further includes the steps of: two images to be matched are obtained by a feature extractor
Figure FDA0003441409910000013
Figure FDA0003441409910000014
And
Figure FDA0003441409910000015
where d is the dimension of the feature vector, n1And n2Respectively counting the number of the characteristic points of the two images; f1And F2The outputs extracted from the relu4_2 and relu5_1 layers of the VGG-16 neural network are spliced.
3. The graph matching method for homogeneous image retrieval according to claim 1, wherein said step S4 further includes the steps of: the attribute of each edge is formed by two end point coordinates after normalization, and the connection information of the edge represents the topological structure information of each graph; then, inputting the point characteristic information and the edge attribute information into a neural network SplineCNN of the graph as input information; SplineCNN is used as a geometric topological information embedding technology, and MAX aggregation is adopted in structural information aggregation; finally, point characteristics embedded with respective geometric topological information are obtained
Figure FDA0003441409910000021
And
Figure FDA0003441409910000022
4. the graph matching method for homogeneous image retrieval according to claim 1, wherein said step S5 further includes the steps of: the point-edge correlation matrices of the two graphs are respectively
Figure FDA0003441409910000023
Figure FDA0003441409910000024
And
Figure FDA0003441409910000025
wherein e1And e2Respectively representing the number of edges of the two figures, when Gi,k=Hj,kWhen 1, it indicates that the edge k starts from node i to node j ends; characteristics of the edge
Figure FDA0003441409910000026
Figure FDA0003441409910000027
And
Figure FDA0003441409910000028
is defined as follows:
Figure FDA0003441409910000029
5. the graph matching method for homogeneous image retrieval according to claim 1, wherein said step S6 further includes the steps of: edge-to-edge recognition matrix Ke
Figure FDA00034414099100000210
Wherein the content of the first and second substances,
Figure FDA00034414099100000211
is a training parameter; keEach element of the matrix represents edge-to-edge matching information, and in order to expand the difference between the similarity values of the edges, i.e. emphasize the high similarity values and compress the low similarity values, the Ke matrix is normalized to obtain a normalized epsilon matrix:
ε=softmax(Ke) Formula (3)
Then, the normalized epsilon matrix is converted into a cross-edge conversion matrix through the structure of the accompanying graph
Figure FDA00034414099100000212
Figure FDA00034414099100000213
Conversion matrix based on cross-graph
Figure FDA00034414099100000214
We can get the feature embedding information across the graph; for node
Figure FDA00034414099100000215
Figure FDA00034414099100000216
Cross-map feature information mj→iIs calculated as follows:
Figure FDA00034414099100000217
finally, vector addition operation is carried out on the cross-map feature information and the point feature information:
Figure FDA00034414099100000218
similar operations are also performed for the feature points of the second graph.
6. The graph matching method for homogeneous image retrieval according to claim 1, wherein said step S7 further includes the steps of: the similarity matrix formula is as follows:
Figure FDA00034414099100000219
the linear solution of the graph matching problem depends on a Sinkhorn iterative algorithm, and the algorithm is sequentially normalized along rows and columns according to the obtained fractional matrix S to obtain a soft distribution matrix
Figure FDA0003441409910000031
Pij=Sinkhorn(exp(Sij) Equation (8).
7. The graph matching method for homogeneous image retrieval according to claim 1, further comprising step S8: given truth allocation matrix
Figure FDA0003441409910000032
And a soft distribution matrix P, wherein the error can be obtained by constructing a cross entropy loss function:
Figure FDA0003441409910000033
8. the graph matching method for homogeneous image retrieval according to claim 1, wherein said step S1 further includes the steps of: the data set contains several different categories of images: airplanes, bicycles, birds, boats, bottles, buses, cars, cats, chairs, cattle, tables, dogs, horses, motorcycles, people, plants, sheep, sofas, trains, televisions; each image contains 6 to 23 annotated feature point image coordinates.
9. The map matching method for homogeneous image retrieval according to claim 1, wherein said step S2 further includes the steps of: 1682 sheets were selected accordingly as test sets. For each image to be trained, extracting a bounding box containing all annotation feature points, adjusting the size of the image to be 256 multiplied by 256, and finally entering the deep learning network training.
CN202111634430.5A 2021-12-29 2021-12-29 Picture matching method for similar image retrieval Active CN114491122B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111634430.5A CN114491122B (en) 2021-12-29 2021-12-29 Picture matching method for similar image retrieval

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111634430.5A CN114491122B (en) 2021-12-29 2021-12-29 Picture matching method for similar image retrieval

Publications (2)

Publication Number Publication Date
CN114491122A true CN114491122A (en) 2022-05-13
CN114491122B CN114491122B (en) 2023-07-14

Family

ID=81496804

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111634430.5A Active CN114491122B (en) 2021-12-29 2021-12-29 Picture matching method for similar image retrieval

Country Status (1)

Country Link
CN (1) CN114491122B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115063789A (en) * 2022-05-24 2022-09-16 中国科学院自动化研究所 3D target detection method and device based on key point matching

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106126581A (en) * 2016-06-20 2016-11-16 复旦大学 Cartographical sketching image search method based on degree of depth study
CN108595636A (en) * 2018-04-25 2018-09-28 复旦大学 The image search method of cartographical sketching based on depth cross-module state correlation study
CN110263795A (en) * 2019-06-04 2019-09-20 华东师范大学 One kind is based on implicit shape and schemes matched object detection method
CN111488498A (en) * 2020-04-08 2020-08-04 浙江大学 Node-graph cross-layer graph matching method and system based on graph neural network
CN112801206A (en) * 2021-02-23 2021-05-14 中国科学院自动化研究所 Image key point matching method based on depth map embedded network and structure self-learning

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106126581A (en) * 2016-06-20 2016-11-16 复旦大学 Cartographical sketching image search method based on degree of depth study
CN108595636A (en) * 2018-04-25 2018-09-28 复旦大学 The image search method of cartographical sketching based on depth cross-module state correlation study
CN110263795A (en) * 2019-06-04 2019-09-20 华东师范大学 One kind is based on implicit shape and schemes matched object detection method
CN111488498A (en) * 2020-04-08 2020-08-04 浙江大学 Node-graph cross-layer graph matching method and system based on graph neural network
CN112801206A (en) * 2021-02-23 2021-05-14 中国科学院自动化研究所 Image key point matching method based on depth map embedded network and structure self-learning

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
ZHOU F等: "Factorized graph matching", 《IEEE TRANSACTIONS ON PATTERN》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115063789A (en) * 2022-05-24 2022-09-16 中国科学院自动化研究所 3D target detection method and device based on key point matching
CN115063789B (en) * 2022-05-24 2023-08-04 中国科学院自动化研究所 3D target detection method and device based on key point matching

Also Published As

Publication number Publication date
CN114491122B (en) 2023-07-14

Similar Documents

Publication Publication Date Title
CN111753060B (en) Information retrieval method, apparatus, device and computer readable storage medium
CN110232152B (en) Content recommendation method, device, server and storage medium
CN109829104B (en) Semantic similarity based pseudo-correlation feedback model information retrieval method and system
US7437349B2 (en) Adaptive probabilistic query expansion
CN108268600B (en) AI-based unstructured data management method and device
CN111353030A (en) Knowledge question and answer retrieval method and device based on travel field knowledge graph
CN110765277B (en) Knowledge-graph-based mobile terminal online equipment fault diagnosis method
CN110399515B (en) Picture retrieval method, device and system
CN106033426B (en) Image retrieval method based on latent semantic minimum hash
Shi et al. Deep adaptively-enhanced hashing with discriminative similarity guidance for unsupervised cross-modal retrieval
CN111985228B (en) Text keyword extraction method, text keyword extraction device, computer equipment and storage medium
CN109408578B (en) Monitoring data fusion method for heterogeneous environment
CN108509543B (en) Streaming RDF data multi-keyword parallel search method based on Spark Streaming
CN115796181A (en) Text relation extraction method for chemical field
CN114911951A (en) Knowledge graph construction method for man-machine cooperation assembly task
CN111325030A (en) Text label construction method and device, computer equipment and storage medium
CN110866129A (en) Cross-media retrieval method based on cross-media uniform characterization model
CN114491122B (en) Picture matching method for similar image retrieval
CN113449066B (en) Method, processor and storage medium for storing cultural relic data by using knowledge graph
Prasomphan Toward Fine-grained Image Retrieval with Adaptive Deep Learning for Cultural Heritage Image.
CN116756347A (en) Semantic information retrieval method based on big data
CN116523041A (en) Knowledge graph construction method, retrieval method and system for equipment field and electronic equipment
CN115797795A (en) Remote sensing image question-answering type retrieval system and method based on reinforcement learning
CN111506754B (en) Picture retrieval method, device, storage medium and processor
CN114036946B (en) Text feature extraction and auxiliary retrieval system and method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant