CN111368125A - Distance measurement method for image retrieval - Google Patents

Distance measurement method for image retrieval Download PDF

Info

Publication number
CN111368125A
CN111368125A CN202010149888.0A CN202010149888A CN111368125A CN 111368125 A CN111368125 A CN 111368125A CN 202010149888 A CN202010149888 A CN 202010149888A CN 111368125 A CN111368125 A CN 111368125A
Authority
CN
China
Prior art keywords
image
hausdorff distance
cost function
sift
sift descriptor
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010149888.0A
Other languages
Chinese (zh)
Other versions
CN111368125B (en
Inventor
孙晓明
张宁
车畅
刘野
吴海滨
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Harbin University of Science and Technology
Original Assignee
Harbin University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harbin University of Science and Technology filed Critical Harbin University of Science and Technology
Priority to CN202010149888.0A priority Critical patent/CN111368125B/en
Publication of CN111368125A publication Critical patent/CN111368125A/en
Application granted granted Critical
Publication of CN111368125B publication Critical patent/CN111368125B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/55Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • G06F18/232Non-hierarchical techniques
    • G06F18/2321Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
    • G06F18/23213Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions with fixed number of clusters, e.g. K-means clustering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Databases & Information Systems (AREA)
  • Library & Information Science (AREA)
  • Multimedia (AREA)
  • Probability & Statistics with Applications (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a distance measurement method facing image retrieval, which comprises the following steps of improving the traditional Hausdorff distance measurement: firstly, writing a differential equation form of a cost function, then obtaining a general solution of the cost function, and finally, taking the traditional Hausdorff distance as a variable of the cost function to improve the Hausdorff distance; the distance measurement method is applied to an image retrieval method based on combination of vocabulary tree information fusion and Hausdorff distance, can improve the image retrieval accuracy and provide a theoretical basis for image retrieval with a complex background.

Description

Distance measurement method for image retrieval
The application is a divisional application of the invention patent application, namely 'image retrieval method based on combination of vocabulary tree information fusion and Hausdorff distance'.
Application date of the original case: 2017-02-13.
Original application No.: 2017100760427.
the name of the original invention is: an image retrieval method based on combination of vocabulary tree information fusion and Hausdorff distance.
Technical Field
The invention discloses a distance measurement method for image retrieval, belongs to the technical field of image retrieval, and particularly relates to a key step in an image retrieval method based on combination of vocabulary tree information fusion and Hausdorff distance.
Background
Image retrieval methods have so far formed three important branches: text-based image retrieval, content-based image retrieval, and semantic-based image retrieval.
The image retrieval based on the text describes the requirements of users by using texts such as image names, image characteristics and the like, but because the text expression capacity is limited and the text annotation has ambiguity, the retrieval result is often inconsistent with the requirements of the users;
the semantic-based image retrieval further refines the high-level semantic expression capability of the image on the basis of the visual characteristics of the image, but the retrieval process of the retrieval method is complex, and the problem that the method system is not developed completely exists;
the image retrieval based on the content is carried out by taking color, texture, shape and the like as the characteristic expression of the image and taking the characteristic expression as the basis for judging the similarity.
If the image features can be extracted accurately, content-based image retrieval would have accuracy advantages not available with the other two types of retrieval. The technical advantages are also aimed at by broad scholars, and researches on how to improve the accuracy of image feature extraction are carried out, so that the accuracy of content-based image retrieval is expected to be further improved.
Disclosure of Invention
In order to meet the technical requirements, the invention discloses an image retrieval method based on combination of vocabulary tree information fusion and Hausdorff distance, which can effectively improve the accuracy of content-based image retrieval.
The purpose of the invention is realized as follows:
the image retrieval method based on combination of vocabulary tree information fusion and Hausdorff distance comprises the following steps:
step a, extracting an image to be retrieved and SIFT characteristics of an image library;
b, generating an SIFT descriptor histogram and SIFT descriptor kernel density;
step c, fusing SIFT descriptor kernel density and SIFT descriptor histogram;
step d, improving the traditional Hausdorff distance measurement;
and e, using the improved Hausdorff distance for image matching.
The image retrieval method based on combination of vocabulary tree information fusion and Hausdorff distance comprises the following specific steps of the step a:
step a 1: constructing to-be-retrieved image and image library Gaussian difference scale function
The method comprises the following steps of performing convolution operation on Gaussian functions with different scales and an image to construct a Gaussian difference scale function D (x, y, sigma) of a two-dimensional image, wherein the method comprises the following steps:
D(x,y,σ)=(G(x,y,kσ)-G(x,y,σ))*I(x,y)
where k is the scale scaling factor, G (x, y, σ) is a gaussian function of variable scale, I (x, y) is the image, and there are:
Figure BDA0002402052840000021
wherein, (x, y) is a scale coordinate, and the size of σ determines the degree of image smoothness;
step a 2: detecting extreme point in Gaussian difference scale space
Comparing each sampling point in the image with adjacent points of the sampling point, and when a certain sampling point is the maximum value or the minimum value in all points of the Gaussian difference scale space, considering the certain sampling point as a characteristic point of the image under the scale;
step a 3: removing the feature points with unstable edges to generate SIFT descriptors
And removing unstable characteristic points of the edge by using a Harris Corner detector, and keeping stable characteristic points to generate the SIFT descriptor.
The image retrieval method based on the combination of the vocabulary tree information fusion and the Hausdorff distance comprises the following specific steps of the step b:
step b 1: construction of an extensible lexical tree by hierarchical clustering of SIFT descriptors
Extracting SIFT descriptors of each picture to obtain a set F ═ FiThen, the set F is subjected to hierarchical clustering by adopting a K-Means clustering method, and initially, the set F is subjected to K-Means clustering at the 1 st layer, and is divided into K parts of { F }iI is more than or equal to 1 and less than or equal to k; repeating the operation until the depth reaches the preset L value to construct an expandable vocabulary tree, wherein c is B in totalLA node, wherein B is a branching factor, L is a depth, c is a total number of nodes, fiRepresenting a certain SIFT descriptor in a picture, F being a set of descriptors, FiIs a certain cluster set obtained by carrying out K-Means clustering on the set F;
step b 2: accumulating the times of the occurrence of the descriptors on each node in the extensible vocabulary tree to obtain an SIFT descriptor histogram
In constructing an expandable lexical tree, c is shared as BLEach node accumulates the occurrence frequency of SIFT descriptors on the first node to obtain SIFT descriptor histogram based on the expandable vocabulary tree, and H is used as [ H ]1,...,hi,...,hc]Is represented by the formula (I) in which hiRepresenting the times of SIFT descriptors appearing on the ith node;
step b 3: quantizing the SIFT descriptor to obtain SIFT descriptor kernel density
Quantizing all SIFT descriptors, then each SIFT descriptor fiAll corresponding to a quantization path from a root node to a leaf node in the expandable vocabulary tree, i.e. corresponding to a group of visual words
Figure BDA0002402052840000031
Each group of visual words corresponds to the kernel density f (c) thereof, and the SIFT descriptor kernel density based on the extensible vocabulary tree is obtained; wherein
Figure BDA0002402052840000032
Is a visual word, i.e. each node in the expandable vocabulary tree represents a visual word, l represents the number of layers of the node in the expandable vocabulary tree, hlIndicating the index of the node in the level tree node, L being the depth.
The image retrieval method based on the combination of the vocabulary tree information fusion and the Hausdorff distance comprises the following specific steps of step c:
step c 1: obtaining a basic probability distribution function of SIFT descriptor histogram and SIFT descriptor kernel density
For computational convenience, if the SIFT descriptor histogram is set as a, and the SIFT descriptor kernel density is set as B, then the box Ω: { A, B }, the decision box is a set describing all elements constituting the whole hypothesis space, and is represented by m () with all possible results considered by the basic probability distribution function; at this time, the process of the present invention,
the basic probability distribution function of subset A is
Figure BDA0002402052840000033
The basic probability distribution function of the subset B is
Figure BDA0002402052840000034
Wherein, M is a normalization constant,
Figure BDA0002402052840000035
m1(Ai) Denotes that the focal length is AiBasic probability assignment of (1), m2(Bj) Representing the focal unitIs BjAssigning a basic probability;
step c 2: the fusion result is obtained by applying Dempster combination rule and combining the step c1
The Dempster combination rule is:
Figure BDA0002402052840000036
substituting the results m (A) and m (B) obtained in the step c1 into m (AB);
wherein M is a normalization constant, and M is ∑A∩B=φ(m(A)m(B))=1-∑A∩B≠φ(m(A)m(B))
m (a) represents the basic probability distribution function of subset a, m (B) represents the basic probability distribution function of subset B, and m (ab) represents the fused basic probability distribution function of subset a and subset B.
The image retrieval method based on the combination of the vocabulary tree information fusion and the Hausdorff distance comprises the following specific steps of the step d:
step d 1: form of differential equation for writing cost function
The differential equation form of the cost function is as follows:
Figure BDA0002402052840000041
step d 2: obtaining a general solution to a cost function
Solving the differential equation to obtain the cost function with the following expression:
Figure BDA0002402052840000042
wherein gamma is0The cost function is an initial value of the cost function, the range of the cost function is 0-1, k is a proportionality coefficient, and tau is a matching parameter;
step d 3: improved Hausdorff distance using traditional Hausdorff distance as variable of cost function
Given two finite sets X ═ X1,x2,...,xMY ═ Y1,y2,...,yNThe conventional Hausdorff distance between X and Y is defined as
Figure BDA0002402052840000043
Where d (X, Y) is the conventional Hausdorff distance, min represents the minimum, max represents the maximum, X and Y are the points in the point sets X and Y, respectively, and d (X, Y) represents the geometric distance between point X and point Y;
the improved Hausdorff distance is:
Figure BDA0002402052840000044
where | X | is the number of the finite set X, dH(X, Y) is the modified Hausdorff distance, d (X, Y) is the conventional Hausdorff distance, and γ (d (X, Y)) is a cost function with a variable d (X, Y).
The image retrieval method based on combination of vocabulary tree information fusion and Hausdorff distance comprises the following specific steps of step e:
and c, according to the fusion characteristics obtained in the step c, carrying out image similarity measurement by using an improved Hausdorff distance, and arranging the obtained similarities according to a descending order to obtain a retrieval result.
Has the advantages that:
the method adopts the following technical means that firstly, SIFT features of an image to be retrieved and an image library are extracted, then an SIFT descriptor histogram and SIFT descriptor kernel density are generated, then the SIFT descriptor kernel density and the SIFT descriptor histogram are fused, the traditional Hausdorff distance measurement is improved, and finally the improved Hausdorff distance is used for image matching; the technical means are interdependent and are all absent, and as a whole, the technical purpose that the content-based image retrieval accuracy rate cannot be effectively improved when any one means exists is achieved.
Drawings
FIG. 1 is a flow chart of the image retrieval method based on the combination of lexical tree information fusion and Housdov distance.
FIG. 2 is a graph comparing precision ratios of three methods.
Fig. 3 is a "banyan" image to be retrieved.
Fig. 4 is a "banyan" search result based on the method of the present invention.
Fig. 5 is a "banyan" search result based on the SIFT descriptor histogram method.
Fig. 6 is a "banyan" search result based on the SIFT descriptor kernel density method.
FIG. 7 is a "tiger" image to be retrieved.
FIG. 8 is the "tiger" search result based on the method of the present invention.
Fig. 9 is a "tiger" search result based on the SIFT descriptor histogram method.
Fig. 10 is a "tiger" search result based on the SIFT descriptor kernel density method.
Detailed Description
The following describes embodiments of the present invention in further detail with reference to the accompanying drawings.
Detailed description of the preferred embodiment
The present embodiment is a theoretical embodiment of an image retrieval method based on the combination of lexical tree information fusion and hausdorff distance.
The image retrieval method based on the combination of vocabulary tree information fusion and Hausdorff distance in the embodiment has a flow chart as shown in FIG. 1, and comprises the following steps:
step a, extracting an image to be retrieved and SIFT characteristics of an image library;
b, generating an SIFT descriptor histogram and SIFT descriptor kernel density;
step c, fusing SIFT descriptor kernel density and SIFT descriptor histogram;
step d, improving the traditional Hausdorff distance measurement;
and e, using the improved Hausdorff distance for image matching.
The image retrieval method based on combination of vocabulary tree information fusion and Hausdorff distance comprises the following specific steps of the step a:
step a 1: constructing to-be-retrieved image and image library Gaussian difference scale function
The method comprises the following steps of performing convolution operation on Gaussian functions with different scales and an image to construct a Gaussian difference scale function D (x, y, sigma) of a two-dimensional image, wherein the method comprises the following steps:
D(x,y,σ)=(G(x,y,kσ)-G(x,y,σ))*I(x,y)
where k is the scale scaling factor, G (x, y, σ) is a gaussian function of variable scale, I (x, y) is the image, and there are:
Figure BDA0002402052840000061
wherein, (x, y) is a scale coordinate, and the size of σ determines the degree of image smoothness;
step a 2: detecting extreme point in Gaussian difference scale space
Comparing each sampling point in the image with adjacent points of the sampling point, and when a certain sampling point is the maximum value or the minimum value in all points of the Gaussian difference scale space, considering the certain sampling point as a characteristic point of the image under the scale;
step a 3: removing the feature points with unstable edges to generate SIFT descriptors
And removing unstable characteristic points of the edge by using a Harris Corner detector, and keeping stable characteristic points to generate the SIFT descriptor.
The image retrieval method based on the combination of the vocabulary tree information fusion and the Hausdorff distance comprises the following specific steps of the step b:
step b 1: construction of an extensible lexical tree by hierarchical clustering of SIFT descriptors
Extracting SIFT descriptors of each picture to obtain a set F ═ FiThen, the set F is subjected to hierarchical clustering by adopting a K-Means clustering method, and initially, the set F is subjected to K-Means clustering at the 1 st layer, and is divided into K parts of { F }iI is more than or equal to 1 and less than or equal to k; repeating the above operations until the depth reaches the preset L value, constructing an expandable vocabulary tree, and sharingc=BLA node, wherein B is a branching factor, L is a depth, c is a total number of nodes, fiRepresenting a certain SIFT descriptor in a picture, F being a set of descriptors, FiIs a certain cluster set obtained by carrying out K-Means clustering on the set F;
step b 2: accumulating the times of the occurrence of the descriptors on each node in the extensible vocabulary tree to obtain an SIFT descriptor histogram
In constructing an expandable lexical tree, c is shared as BLEach node accumulates the occurrence frequency of SIFT descriptors on the first node to obtain SIFT descriptor histogram based on the expandable vocabulary tree, and H is used as [ H ]1,...,hi,...,hc]Is represented by the formula (I) in which hiRepresenting the times of SIFT descriptors appearing on the ith node;
step b 3: quantizing the SIFT descriptor to obtain SIFT descriptor kernel density
Quantizing all SIFT descriptors, then each SIFT descriptor fiAll corresponding to a quantization path from a root node to a leaf node in the expandable vocabulary tree, i.e. corresponding to a group of visual words
Figure BDA0002402052840000071
Each group of visual words corresponds to the kernel density f (c) thereof, and the SIFT descriptor kernel density based on the extensible vocabulary tree is obtained; wherein
Figure BDA0002402052840000072
Is a visual word, i.e. each node in the expandable vocabulary tree represents a visual word, l represents the number of layers of the node in the expandable vocabulary tree, hlIndicating the index of the node in the level tree node, L being the depth.
The image retrieval method based on the combination of the vocabulary tree information fusion and the Hausdorff distance comprises the following specific steps of step c:
step c 1: obtaining a basic probability distribution function of SIFT descriptor histogram and SIFT descriptor kernel density
For computational convenience, if the SIFT descriptor histogram is set as a, and the SIFT descriptor kernel density is set as B, then the box Ω: { A, B }, the decision box is a set describing all elements constituting the whole hypothesis space, and is represented by m () with all possible results considered by the basic probability distribution function; at this time, the process of the present invention,
the basic probability distribution function of subset A is
Figure BDA0002402052840000073
The basic probability distribution function of the subset B is
Figure BDA0002402052840000074
Wherein, M is a normalization constant,
Figure BDA0002402052840000075
m1(Ai) Denotes that the focal length is AiBasic probability assignment of (1), m2(Bj) Denotes that the focal length is BjAssigning a basic probability;
step c 2: the fusion result is obtained by applying Dempster combination rule and combining the step c1
The Dempster combination rule is:
Figure BDA0002402052840000076
substituting the results m (A) and m (B) obtained in the step c1 into m (AB);
wherein M is a normalization constant, and M is ∑A∩B=φ(m(A)m(B))=1-∑A∩B≠φ(m(A)m(B))
m (a) represents the basic probability distribution function of subset a, m (B) represents the basic probability distribution function of subset B, and m (ab) represents the fused basic probability distribution function of subset a and subset B.
The image retrieval method based on the combination of the vocabulary tree information fusion and the Hausdorff distance comprises the following specific steps of the step d:
step d 1: form of differential equation for writing cost function
The differential equation form of the cost function is as follows:
Figure BDA0002402052840000081
step d 2: obtaining a general solution to a cost function
Solving the differential equation to obtain the cost function with the following expression:
Figure BDA0002402052840000082
wherein gamma is0The cost function is an initial value of the cost function, the range of the cost function is 0-1, k is a proportionality coefficient, and tau is a matching parameter;
step d 3: improved Hausdorff distance using traditional Hausdorff distance as variable of cost function
Given two finite sets X ═ X1,x2,...,xMY ═ Y1,y2,...,yNThe conventional Hausdorff distance between X and Y is defined as
Figure BDA0002402052840000083
Where d (X, Y) is the conventional Hausdorff distance, min represents the minimum, max represents the maximum, X and Y are the points in the point sets X and Y, respectively, and d (X, Y) represents the geometric distance between point X and point Y;
the improved Hausdorff distance is:
Figure BDA0002402052840000084
where | X | is the number of the finite set X, dH(X, Y) is the modified Hausdorff distance, d (X, Y) is the conventional Hausdorff distance, and γ (d (X, Y)) is a cost function with a variable d (X, Y).
The image retrieval method based on combination of vocabulary tree information fusion and Hausdorff distance comprises the following specific steps of step e:
and c, according to the fusion characteristics obtained in the step c, carrying out image similarity measurement by using an improved Hausdorff distance, and arranging the obtained similarities according to a descending order to obtain a retrieval result.
Detailed description of the invention
The present embodiment is a theoretical embodiment of an image retrieval method based on the combination of lexical tree information fusion and hausdorff distance.
In view of the fact that most of the technicians in the field are scholars, the writing of technical documents is more customary to the writing of articles, and therefore, on the basis of no essential difference from the specific embodiment, the second specific embodiment is supplemented according to the scholars.
The image retrieval method based on combination of vocabulary tree information fusion and Hausdorff distance in the embodiment comprises the following steps:
step a: SIFT feature extraction (SIFT: scale invariant feature transform) of image to be retrieved and image library
Step a 1: constructing to-be-retrieved image and image library Gaussian difference scale function
During the extraction of the SIFT descriptor, a gaussian difference scale space is firstly constructed, the scale space of the two-dimensional image is L (x, y, sigma) · G (x, y, sigma) · I (x, y),
Figure BDA0002402052840000091
where G (x, y, σ) is a gaussian function with variable scale, (x, y) is the scale coordinate, I (x, y) is the image, L (x, y, σ) is the scale space of the two-dimensional image, and the size of σ determines the degree of smoothing of the image.
For more accurate detection of image feature points, it is necessary to construct a gaussian difference scale function of a two-dimensional image, which is generated by convolving the image with gaussian functions of different scales, i.e., D (x, y, σ) — (G (x, y, k σ) -G (x, y, σ)) × I (x, y) — L (x, y, k σ) -L (x, y, σ), where D (x, y, σ) is the gaussian difference scale function of the two-dimensional image and k is a scale scaling factor
Step a 2: detecting extreme point in Gaussian difference scale space
To find the extreme point in the scale space requires that each sample point in the image is compared with its neighboring points, and when a sample point is the maximum or minimum among all points in the DoG (difference of gaussians) space, the point is considered as a feature point of the image under the scale.
Step a 3: removing the feature points with unstable edges to generate SIFT descriptors
In order to enhance the matching stable point and improve the noise capability, a Harris Commer detector is used for removing unstable characteristic points of the edge. And keeping stable feature points to generate SIFT descriptors.
Step b: generating SIFT descriptor histogram and SIFT descriptor kernel density
Step b 1: construction of an extensible lexical tree SVT by hierarchical clustering of SIFT descriptors
Extracting SIFT descriptors of each picture to obtain a set F ═ FiAnd then, performing hierarchical clustering on the set F by adopting a K-Means clustering method. Initially, K-Means clustering is performed on the set F at layer 1, and the set F is divided into K parts of { FiI is more than or equal to 1 and less than or equal to k. Similarly, the newly generated cluster is subdivided into K clusters by using K-Means, and the above operations are repeated until the depth reaches the preset L value, so that the split is not performed, and an expandable vocabulary tree is constructed, wherein c is BLAnd each node is formed. Where B is the branching factor, L is the depth, c is the total number of nodes, fiRepresenting a certain SIFT descriptor in a picture, F being a set of descriptors, FiIs a certain cluster set obtained by performing K-Means clustering on the set F.
Step b 2: accumulating the times of the occurrence of the descriptors on each node in the extensible vocabulary tree to obtain an SIFT descriptor histogram
In constructing an expandable lexical tree, c is shared as BLEach node accumulates the occurrence frequency of SIFT descriptors on the first node to obtain H [ H ] for SIFT descriptor histogram based on the expandable vocabulary tree1,...,hi,...,hc]Is represented by the formula (I) in which hiThe number of times of SIFT descriptors appearing at the ith node is represented, B is a branching factor, L is depth, and c is the total number of nodes.
Step b 3: quantizing the SIFT descriptor to obtain SIFT descriptor kernel density
Quantizing all SIFT descriptors, then each SIFT descriptor fiAll correspond to a subordinate root section in the expandable vocabulary treeQuantization paths from point to leaf node, i.e. corresponding to a set of visual words
Figure BDA0002402052840000101
Each group of visual words corresponds to the kernel density f (c), and the SIFT descriptor kernel density based on the extensible vocabulary tree is obtained. Wherein
Figure BDA0002402052840000102
Is a visual word, i.e. each node in the expandable vocabulary tree represents a visual word, l represents the number of layers of the node in the expandable vocabulary tree, hlIndicating the index of the node in the level tree node, L being the depth.
Step c: fusing SIFT descriptor kernel density and SIFT descriptor histogram
Step c 1: obtaining a basic probability distribution function of SIFT descriptor histogram and SIFT descriptor kernel density
For the following computational convenience, the frame Ω is identified by setting the SIFT descriptor histogram to a and the SIFT descriptor kernel density to B: { A, B }, the discrimination box is a set of all elements describing the overall hypothetical space. All possible outcomes are considered with the basic probability distribution function (BPA), often denoted m ().
The basic probability distribution function of subset A is
Figure BDA0002402052840000103
The basic probability distribution function of the subset B is
Figure BDA0002402052840000104
Wherein, M is a normalization constant,
Figure BDA0002402052840000105
m1(Ai) Denotes that the focal length is AiBasic probability assignment of (1), m2(Bj) Denotes that the focal length is BjAssigning a basic probability;
step c 2: the fusion result is obtained by applying Dempster combination rule and combining the step c1
The Dempster combination rule is:
Figure BDA0002402052840000106
substituting the results m (A) and m (B) obtained in step c1 into m (AB).
Wherein M is a normalization constant, and M is ∑A∩B=φ(m(A)m(B))=1-∑A∩B≠φ(m(A)m(B))
m (a) represents the basic probability distribution function of subset a, m (B) represents the basic probability distribution function of subset B, and m (ab) represents the fused basic probability distribution function of subset a and subset B.
Step d: improving the conventional Hausdorff distance metric
In order to improve the reliability and stability of the matching process, the invention improves the traditional Hausdorff distance measurement, namely, the traditional Hausdorff distance is used as a variable of a cost function as the improved Hausdorff distance.
Step d 1: form of differential equation for writing cost function
The differential equation form of the cost function is as follows:
Figure BDA0002402052840000111
step d 2: obtaining a general solution to a cost function
Solving the differential equation to obtain the cost function with the following expression:
Figure BDA0002402052840000112
wherein gamma is0The cost function is an initial value of the cost function, the range of the cost function is 0-1, k is a proportionality coefficient, and tau is a matching parameter.
Step d 3: improved Hausdorff distance using traditional Hausdorff distance as variable of cost function
Given two finite sets X ═ X1,x2,...,xMY ═ Y1,y2,...,yNThe conventional Hausdorff distance between X and Y is defined as
Figure BDA0002402052840000113
Where d (X, Y) is the conventional Hausdorff distance, min represents the minimum, max represents the maximum, X and Y are the points in the point sets X and Y, respectively, and d (X, Y) represents the geometric distance between point X and point Y
The improved Hausdorff distance is:
Figure BDA0002402052840000114
where | X | is the number of the finite set X, dH(X, Y) is the modified Hausdorff distance, d (X, Y) is the conventional Hausdorff distance, and γ (d (X, Y)) is a cost function with a variable d (X, Y)
Step e: using improved Hausdorff distance for image matching
And c, according to the fusion characteristics obtained in the step c, carrying out image similarity measurement on the characteristics by using an improved Hausdorff distance, and arranging the obtained similarities according to a descending order to obtain a retrieval result.
Detailed description of the preferred embodiment
The present embodiment is an experimental embodiment of an image retrieval method based on combination of lexical tree information fusion and hausdov distance.
Fig. 2 shows precision rates of image retrieval based on the SIFT descriptor histogram, image retrieval based on the SIFT descriptor kernel density, and image retrieval based on the present invention.
As can be seen from fig. 2, the first four cloud, star, bird and tree in the image category are simple pictures with backgrounds, and the precision ratios of the three retrieved images are not very different; the last four items in the image category are pictures with tiger, fish, mountain and flower as complex backgrounds, the search accuracy of the three search methods is greatly different, and the search of the invention is far larger than the search of the first two.
The experimental results for the two image types are given below
In the experiment, a small self-built image database is used, wherein the database contains 8 types of images, namely flowers, birds, fish, tigers, mountains, trees, stars and clouds, the total number of the images is 800 in total, and each type of image is 100.
Experiment one: background clearness experiment of image to be retrieved
The method comprises the steps of taking a banyan image with a simple background as an image to be retrieved, randomly extracting 5 images from all banyans as query images, and finally selecting an average value of the precision ratios of the 5 images as a final result. The precision ratio is defined as follows: the precision ratio (number of images related to the key map in the query result/number of images returned by the query) is 100%.
A banyan image with a simple background is given as an image to be retrieved, as shown in fig. 3; the retrieval result of the method of the invention is shown in fig. 4, the retrieval result based on the SIFT descriptor histogram method is shown in fig. 5, and the retrieval result based on the SIFT descriptor kernel density method is shown in fig. 6.
As can be seen from the search results of fig. 4, 5, and 6: the background of the image to be retrieved is clear, the color information of the banyan is clear, the crown of the banyan is large, most of the image is covered, and rich textural feature information is formed; the shape information between the crown and the background of the image to be retrieved and at the trunk is clearer.
Each image to be retrieved returns 30 images, wherein the images accurately retrieved by the method of the invention are respectively 23, 25 and 25, the precision ratio is respectively 76.7%, 83.3% and 83.3%, and the average precision ratio (76.7+76.7+83.3+83.3+83.3)/5 is 100% and 80.66%; the images accurately searched by the SIFT descriptor histogram method are 23, 24, 25 and 25 respectively, the precision ratio is 76.7%, 80%, 83.3% and 83.3%, and the average precision ratio is (76.7+76.7+80+83.3+83.3)/5 x 100% is 80%; the images accurately searched by the SIFT descriptor kernel density method are 23, 24, 25 and 25 respectively, the precision ratio is 76.7%, 80%, 83.3% and 83.3%, and the average precision ratio is (80+76.7+76.7+83.3+83.3)/5 x 100% is 80%;
for pictures with simple backgrounds, the difference between the retrieval method disclosed by the invention and the images retrieved by the histogram retrieval based on the SIFT descriptor and the kernel density based on the SIFT descriptor is not large, and the precision rate difference is not large and reaches about 80%.
Experiment two: background complex experiment of image to be retrieved
Taking the 'tiger' image with complex background as the image to be retrieved, randomly extracting 5 images from all the 'tiger' images as query images, and finally taking the average value of the precision ratios of the selected 5 images as the final result. The precision ratio is defined as follows: the precision ratio (number of images related to the key map in the query result/number of images returned by the query) is 100%.
A tiger image with a complex background is given as an image to be retrieved, as shown in fig. 7; the retrieval result of the method according to the invention is shown in fig. 8, the retrieval result based on the SIFT descriptor histogram method is shown in fig. 9, and the retrieval result based on the SIFT descriptor kernel density method is shown in fig. 10.
As can be seen from fig. 8, a total of 30 images were returned, of which 26 were accurately retrieved with an accuracy of 86.7%. The first image of the retrieval result is the image to be retrieved, 25 retrieved images in the remaining 29 images are all the images of the tiger class, and the shape of the tiger head, the pattern of the tiger skin, the characteristics of the background area and the like in the 25 images are very similar to the image to be retrieved.
As can be seen from fig. 9, a total of 30 images were returned, of which 12 images were retrieved with an accuracy of 40%. As can be seen from fig. 10, a total of 30 images were returned, of which 13 images were accurately retrieved with an accuracy of 43.3%. The two search results show that although 12 and 13 searched images are also the images of the tiger type, the shape of the tiger head, the patterns of the tiger skin and the background area are greatly different from the image to be searched, but the characteristic background of the searched image is single.
Retrieving the other four images to be retrieved as the images of the tiger, wherein each image to be retrieved returns 30 images, the images accurately retrieved by the method are respectively 25, 26 and 27, the precision ratio is respectively 83.3%, 86.7% and 90.0%, and the average precision ratio is (86.7+83.3+83.3+86.7+90.0)/5 x 100% is 86.0%; the images accurately searched by the SIFT descriptor histogram method are respectively 12, 13 and 13, the precision ratio is respectively 40.0%, 43.3% and 43.3%, and the average precision ratio is (40.0+40.0+40.0+43.3+43.3)/5 is 100% and is 41.32%; the images accurately searched by the SIFT descriptor check density method are respectively 12, 13 and 13, the precision ratio is respectively 40.0%, 40.0% and 43.3%, and the average precision ratio is (43.3+40.0+40.0+43.3+43.3)/5 x 100% is 41.98%;
from the search results of experiment two, it can be found that the average precision ratio of the two search results which are not fused in the search of the pictures with complicated backgrounds only reaches 41.32% and 41.98%, which is equivalent to that the pictures with complicated backgrounds cannot be searched at all. The average precision ratio of the method reaches 86%, and the precision ratio is not reduced due to the complex background, so that the retrieval result fully proves that the image retrieval method combining the expandable vocabulary tree information fusion and the Hausdorff distance can make up the defect that the original retrieval method cannot retrieve the picture with the complex background.

Claims (4)

1. A distance measurement method for image retrieval is characterized in that the traditional Hausdorff distance measurement method is improved, and the method comprises the following steps:
step d 1: form of differential equation for writing cost function
The differential equation form of the cost function is as follows:
Figure FDA0002402052830000011
step d 2: obtaining a general solution to a cost function
Solving the differential equation to obtain the cost function with the following expression:
Figure FDA0002402052830000012
wherein gamma is0The cost function is an initial value of the cost function, the range of the cost function is 0-1, k is a proportionality coefficient, and tau is a matching parameter;
step d 3: improved Hausdorff distance using traditional Hausdorff distance as variable of cost function
Given two finite sets X ═ X1,x2,...,xMY ═ Y1,y2,...,yNThe conventional Hausdorff distance between X and Y is defined as
Figure FDA0002402052830000013
Where d (X, Y) is the conventional Hausdorff distance, min represents the minimum, max represents the maximum, X and Y are the points in the point sets X and Y, respectively, and d (X, Y) represents the geometric distance between point X and point Y;
the improved Hausdorff distance is:
Figure FDA0002402052830000014
where | X | is the number of the finite set X, dH(X, Y) is the modified Hausdorff distance, d (X, Y) is the conventional Hausdorff distance, and γ (d (X, Y)) is a cost function with a variable d (X, Y).
2. The image-search-oriented distance measurement method according to claim 1, used for image search.
3. The image-search-oriented distance measurement method according to claim 2, wherein the image search is an image search method based on vocabulary tree information fusion in combination with hausdorff distance.
4. The image-retrieval-oriented distance measurement method according to claim 3, characterized by comprising the steps of:
step a, extracting an image to be retrieved and SIFT characteristics of an image library; the method comprises the following specific steps:
step a 1: constructing a Gaussian difference scale function of an image to be retrieved and an image library;
step a 2: detecting extreme points in a Gaussian difference scale space;
step a 3: removing feature points with unstable edges and generating SIFT descriptors;
b, generating an SIFT descriptor histogram and SIFT descriptor kernel density; the method comprises the following specific steps:
step b 1: constructing an extensible vocabulary tree through hierarchical clustering of SIFT descriptors;
step b 2: accumulating the occurrence times of the descriptors on each node in the extensible vocabulary tree to obtain an SIFT descriptor histogram;
step b 3: quantizing the SIFT descriptors to obtain SIFT descriptor kernel density;
step c, fusing SIFT descriptor kernel density and SIFT descriptor histogram; the method comprises the following specific steps:
step c 1: obtaining a SIFT descriptor histogram and a basic probability distribution function of SIFT descriptor kernel density;
step c 2: a fusion result is obtained by applying Dempster combination rule and combining the step c 1;
step d, improving the traditional Hausdorff distance measurement; the method comprises the following specific steps:
step d 1: writing a differential equation form of the cost function;
step d 2: obtaining a general solution of the cost function;
step d 3: the traditional Hausdorff distance is used as a variable of the cost function, and the Hausdorff distance is improved;
step e, using the improved Hausdorff distance for image matching; the method comprises the following specific steps:
and c, according to the fusion characteristics obtained in the step c, carrying out image similarity measurement by using an improved Hausdorff distance, and arranging the obtained similarities according to a descending order to obtain a retrieval result.
CN202010149888.0A 2017-02-13 2017-02-13 Distance measurement method for image retrieval Expired - Fee Related CN111368125B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010149888.0A CN111368125B (en) 2017-02-13 2017-02-13 Distance measurement method for image retrieval

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201710076042.7A CN106844733B (en) 2017-02-13 2017-02-13 Image retrieval method based on combination of vocabulary tree information fusion and Hausdorff distance
CN202010149888.0A CN111368125B (en) 2017-02-13 2017-02-13 Distance measurement method for image retrieval

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
CN201710076042.7A Division CN106844733B (en) 2017-02-13 2017-02-13 Image retrieval method based on combination of vocabulary tree information fusion and Hausdorff distance

Publications (2)

Publication Number Publication Date
CN111368125A true CN111368125A (en) 2020-07-03
CN111368125B CN111368125B (en) 2022-06-10

Family

ID=59128893

Family Applications (5)

Application Number Title Priority Date Filing Date
CN202010149888.0A Expired - Fee Related CN111368125B (en) 2017-02-13 2017-02-13 Distance measurement method for image retrieval
CN202010149889.5A Expired - Fee Related CN111309955B (en) 2017-02-13 2017-02-13 Fusion method for image retrieval
CN202010149894.6A Expired - Fee Related CN111309956B (en) 2017-02-13 2017-02-13 Image retrieval-oriented extraction method
CN202010149899.9A Expired - Fee Related CN111368126B (en) 2017-02-13 2017-02-13 Image retrieval-oriented generation method
CN201710076042.7A Expired - Fee Related CN106844733B (en) 2017-02-13 2017-02-13 Image retrieval method based on combination of vocabulary tree information fusion and Hausdorff distance

Family Applications After (4)

Application Number Title Priority Date Filing Date
CN202010149889.5A Expired - Fee Related CN111309955B (en) 2017-02-13 2017-02-13 Fusion method for image retrieval
CN202010149894.6A Expired - Fee Related CN111309956B (en) 2017-02-13 2017-02-13 Image retrieval-oriented extraction method
CN202010149899.9A Expired - Fee Related CN111368126B (en) 2017-02-13 2017-02-13 Image retrieval-oriented generation method
CN201710076042.7A Expired - Fee Related CN106844733B (en) 2017-02-13 2017-02-13 Image retrieval method based on combination of vocabulary tree information fusion and Hausdorff distance

Country Status (1)

Country Link
CN (5) CN111368125B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108009154B (en) * 2017-12-20 2021-01-05 哈尔滨理工大学 Image Chinese description method based on deep learning model
CN109978829B (en) * 2019-02-26 2021-09-28 深圳市华汉伟业科技有限公司 Detection method and system for object to be detected
CN111797268B (en) * 2020-07-17 2023-12-26 中国海洋大学 RGB-D image retrieval method
CN111931791B (en) * 2020-08-11 2022-10-11 重庆邮电大学 Method for realizing image turnover invariance

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040042661A1 (en) * 2002-08-30 2004-03-04 Markus Ulrich Hierarchical component based object recognition
US20050100219A1 (en) * 2003-11-10 2005-05-12 Kathrin Berkner Features for retrieval and similarity matching of documents from the JPEG 2000-compressed domain
WO2008073366A2 (en) * 2006-12-08 2008-06-19 Sobayli, Llc Target object recognition in images and video

Family Cites Families (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5999653A (en) * 1996-01-19 1999-12-07 Xerox Corporation Fast techniques for searching images using the Hausdorff distance
US7542606B2 (en) * 2004-07-29 2009-06-02 Sony Corporation Use of Hausdorff distances in the earth mover linear program
CN100550037C (en) * 2007-11-23 2009-10-14 重庆大学 Utilize and improve Hausdorff apart from the method for extracting the identification human ear characteristic
CN100592297C (en) * 2008-02-22 2010-02-24 南京大学 Multiple meaning digital picture search method based on representation conversion
CN101493891B (en) * 2009-02-27 2011-08-31 天津大学 Characteristic extracting and describing method with mirror plate overturning invariability based on SIFT
WO2011005865A2 (en) * 2009-07-07 2011-01-13 The Johns Hopkins University A system and method for automated disease assessment in capsule endoscopy
US8787682B2 (en) * 2011-03-22 2014-07-22 Nec Laboratories America, Inc. Fast image classification by vocabulary tree based image retrieval
US8811726B2 (en) * 2011-06-02 2014-08-19 Kriegman-Belhumeur Vision Technologies, Llc Method and system for localizing parts of an object in an image for computer vision applications
US20130046793A1 (en) * 2011-08-19 2013-02-21 Qualcomm Incorporated Fast matching of image features using multi-dimensional tree data structures
CN102542058B (en) * 2011-12-29 2013-04-03 天津大学 Hierarchical landmark identification method integrating global visual characteristics and local visual characteristics
CN102662955A (en) * 2012-03-05 2012-09-12 南京航空航天大学 Image retrieval method based on fractal image coding
CN103489176B (en) * 2012-06-13 2016-02-03 中国科学院电子学研究所 A kind of SAR image for serious geometric distortion carries out the method for same place extraction
US8768049B2 (en) * 2012-07-13 2014-07-01 Seiko Epson Corporation Small vein image recognition and authorization using constrained geometrical matching and weighted voting under generic tree model
US9361730B2 (en) * 2012-07-26 2016-06-07 Qualcomm Incorporated Interactions of tangible and augmented reality objects
CN103020111B (en) * 2012-10-29 2015-06-17 苏州大学 Image retrieval method based on vocabulary tree level semantic model
US9177404B2 (en) * 2012-10-31 2015-11-03 Qualcomm Incorporated Systems and methods of merging multiple maps for computer vision based tracking
US8891908B2 (en) * 2012-11-14 2014-11-18 Nec Laboratories America, Inc. Semantic-aware co-indexing for near-duplicate image retrieval
CN102945289B (en) * 2012-11-30 2016-01-06 苏州搜客信息技术有限公司 Based on the image search method of CGCI-SIFT local feature
CN103164856B (en) * 2013-03-07 2014-08-20 南京工业大学 Video copy and paste blind detection method based on dense scale-invariant feature transform stream
CN103336971B (en) * 2013-07-08 2016-08-10 浙江工商大学 Target matching method between multiple-camera based on multiple features fusion and incremental learning
CN103605765B (en) * 2013-11-26 2016-11-16 电子科技大学 A kind of based on the massive image retrieval system clustering compact feature
CN103729654A (en) * 2014-01-22 2014-04-16 青岛新比特电子科技有限公司 Image matching retrieval system on account of improving Scale Invariant Feature Transform (SIFT) algorithm
CN104008174B (en) * 2014-06-04 2017-06-06 北京工业大学 A kind of secret protection index generation method of massive image retrieval
CN104036524A (en) * 2014-06-18 2014-09-10 哈尔滨工程大学 Fast target tracking method with improved SIFT algorithm
CN104915949B (en) * 2015-04-08 2017-09-29 华中科技大学 A kind of image matching method of combination point feature and line feature
CN105183746B (en) * 2015-07-08 2018-04-17 西安交通大学 The method that notable feature realizes image retrieval is excavated from more picture concerneds
CN105022835B (en) * 2015-08-14 2018-01-12 武汉大学 A kind of intelligent perception big data public safety recognition methods and system
CN105138672B (en) * 2015-09-07 2018-08-21 北京工业大学 A kind of image search method of multiple features fusion
CN105550381B (en) * 2016-03-17 2019-04-05 北京工业大学 A kind of efficient image search method based on improvement SIFT feature
CN106294577A (en) * 2016-07-27 2017-01-04 北京小米移动软件有限公司 Figure chip detection method and device
CN106339486A (en) * 2016-08-30 2017-01-18 西安电子科技大学 Image retrieval method based on incremental learning of large vocabulary tree

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040042661A1 (en) * 2002-08-30 2004-03-04 Markus Ulrich Hierarchical component based object recognition
US20050100219A1 (en) * 2003-11-10 2005-05-12 Kathrin Berkner Features for retrieval and similarity matching of documents from the JPEG 2000-compressed domain
CN1617143A (en) * 2003-11-10 2005-05-18 株式会社理光 Features for retrieval and similarity matching of documents
WO2008073366A2 (en) * 2006-12-08 2008-06-19 Sobayli, Llc Target object recognition in images and video
US20080159622A1 (en) * 2006-12-08 2008-07-03 The Nexus Holdings Group, Llc Target object recognition in images and video

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
CHRISTIAN BEECKS: "Signature matching distance for content-based image retrieval", 《PROCEEDINGS OF THE 3RD ACM CONFERENCE ON INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL》, 16 April 2013 (2013-04-16), pages 41 - 48, XP059030708, DOI: 10.1145/2461466.2461474 *
冯晓磊等: "利用梯度方向的Hausdorff距离配准红外和可见光图像", 《电视技术》, no. 03, 2 February 2015 (2015-02-02), pages 11 - 16 *
李倩影: "基于增强现实的人脸卡通化技术研究", 《中国优秀硕士学位论文全文库 信息科技辑》, no. 2, 15 February 2015 (2015-02-15), pages 138 - 823 *
邹鹏等: "基于边界检测的多流形学习算法", 《模式识别与人工智能》, no. 11, 15 November 2016 (2016-11-15), pages 79 - 89 *

Also Published As

Publication number Publication date
CN111309955A (en) 2020-06-19
CN111368125B (en) 2022-06-10
CN111309956A (en) 2020-06-19
CN111309956B (en) 2022-06-24
CN111368126A (en) 2020-07-03
CN106844733B (en) 2020-04-03
CN106844733A (en) 2017-06-13
CN111309955B (en) 2022-06-24
CN111368126B (en) 2022-06-07

Similar Documents

Publication Publication Date Title
CN104850633B (en) A kind of three-dimensional model searching system and method based on the segmentation of cartographical sketching component
Bekkerman et al. Multi-modal clustering for multimedia collections
CN111309955B (en) Fusion method for image retrieval
Fakhari et al. Combination of classification and regression in decision tree for multi-labeling image annotation and retrieval
US20150199567A1 (en) Document classification assisting apparatus, method and program
CN104850822B (en) Leaf identification method under simple background based on multi-feature fusion
Niu et al. Knowledge-based topic model for unsupervised object discovery and localization
CN106874397B (en) Automatic semantic annotation method for Internet of things equipment
CN106874421A (en) Image search method based on self adaptation rectangular window
Pengcheng et al. Fast Chinese calligraphic character recognition with large-scale data
CN105740360B (en) Method for identifying and searching classical titles in artwork images
CN114077682B (en) Intelligent recognition matching processing method and system for image retrieval and storage medium
Ghosh et al. Efficient indexing for query by string text retrieval
EP4127965A1 (en) Computer-implemented method for analogue retrieval of documents
Richter et al. Leveraging community metadata for multimodal image ranking
Pérez-Pimentel et al. A genetic algorithm applied to content-based image retrieval for natural scenes classification
Shi et al. Efficient Image Retrieval via Feature Fusion and Adaptive Weighting
Elhady et al. Weighted feature voting technique for content-based image retrieval
Kumari et al. A Study and usage of Visual Features in Content Based Image Retrieval Systems.
Xu Cross-Media Retrieval: Methodologies and Challenges
Saboorian et al. User adaptive clustering for large image databases
Derakhshan et al. A Review of Methods of Instance-based Automatic Image Annotation
Roman-Rangel Statistical Shape Descriptors for Ancient Maya Hieroglyphs Analysis
Amirshahi Presenting a method based on automatic image annotation techniques for the semantic recovery of images using artificial neural networks
Ni et al. Research on Image Fusion Algorithm Based on Fuzzy Clustering and Semantics

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20220610