CN106557533B - Single-target multi-image joint retrieval method and device - Google Patents

Single-target multi-image joint retrieval method and device Download PDF

Info

Publication number
CN106557533B
CN106557533B CN201510672504.2A CN201510672504A CN106557533B CN 106557533 B CN106557533 B CN 106557533B CN 201510672504 A CN201510672504 A CN 201510672504A CN 106557533 B CN106557533 B CN 106557533B
Authority
CN
China
Prior art keywords
retrieval
target
similarity
feature
features
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201510672504.2A
Other languages
Chinese (zh)
Other versions
CN106557533A (en
Inventor
郭阶添
浦世亮
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Hikvision Digital Technology Co Ltd
Original Assignee
Hangzhou Hikvision Digital Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hangzhou Hikvision Digital Technology Co Ltd filed Critical Hangzhou Hikvision Digital Technology Co Ltd
Publication of CN106557533A publication Critical patent/CN106557533A/en
Application granted granted Critical
Publication of CN106557533B publication Critical patent/CN106557533B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/43Querying
    • G06F16/432Query formulation
    • G06F16/434Query formulation using image data, e.g. images, photos, pictures taken by a user

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a method and a device for single-target multi-image combined retrieval. The method comprises the following steps: receiving a plurality of images comprising the same retrieval target; respectively generating differential weights of the images; fusing the features of the retrieval targets in the multiple images according to the difference weights to generate a feature set; retrieving in an image database based on the feature set to obtain a first retrieval result; and performing weighted fusion on the first retrieval result according to the difference weight, and outputting a weighted and fused second retrieval result. Extracting difference weights and feature sets from a plurality of samples of a retrieval target, retrieving based on the feature sets, and weighting the difference weights into a retrieval result; and a plurality of samples of a single target are fused for retrieval, so that the retrieval result is more visual, and the precision ratio and the recall ratio are higher.

Description

Single-target multi-image joint retrieval method and device
Technical Field
The invention relates to the field of image processing, in particular to a method and a device for single-target multi-image joint retrieval.
Background
The image retrieval is widely applied to the application fields of video monitoring, intelligent analysis, pattern recognition and the like. In the conventional image retrieval, a database is generally searched by extracting a single image to extract features. In a video monitoring scene, the quantity of video data and image data is huge. The images at various angles may be numerous for various lighting conditions, including the same target, that the user grasps. It is a common practice for a user to search one by one and view the search results.
The existing retrieval scheme cannot fuse results of a plurality of query images of one target, and has lower recall ratio for the same target under different illumination conditions and angles.
Disclosure of Invention
The invention aims to provide a method and a device for single-target multi-image joint retrieval, which are characterized in that difference weights and feature sets are extracted from a plurality of samples, retrieval is carried out based on the feature sets, and the difference weights are weighted into retrieval results; and a plurality of samples of a single target are fused for retrieval, so that the retrieval result is more visual, and the precision ratio and the recall ratio are higher.
In order to realize the purpose, the following technical scheme is adopted:
on one hand, the method for single-target multi-image joint retrieval is adopted, and comprises the following steps:
receiving a plurality of images comprising the same retrieval target;
respectively generating differential weights of the images;
fusing the features of the retrieval targets in the multiple images according to the difference weights to generate a feature set;
retrieving in an image database based on the feature set to obtain a first retrieval result;
and performing weighted fusion on the first retrieval result according to the difference weight, and outputting a weighted and fused second retrieval result.
In another aspect, an apparatus for single-target multi-image joint retrieval is adopted, including:
a retrieval target receiving unit for receiving a plurality of images including the same retrieval target;
the difference weight calculation unit is used for respectively generating difference weights of the images;
a feature set generating unit configured to fuse features of the search targets in the plurality of images according to the difference weights to generate a feature set;
the primary retrieval unit is used for retrieving in an image database based on the feature set to obtain a first retrieval result;
and the weighted fusion unit is used for performing weighted fusion on the first retrieval result according to the difference weight and outputting a weighted-fused second retrieval result.
The invention has the beneficial effects that: extracting difference weights and feature sets from a plurality of samples of a retrieval target, retrieving based on the feature sets, and weighting the difference weights into a retrieval result; and a plurality of samples of a single target are fused for retrieval, so that the retrieval result is more visual, and the precision ratio and the recall ratio are higher.
Drawings
Fig. 1 is a flowchart of a method of a first embodiment of a single-target multi-image joint search method according to an embodiment of the present invention.
Fig. 2 is a flowchart of a single-target multi-image joint search method according to a second embodiment of the present invention.
Fig. 3 is a block diagram illustrating a first embodiment of an apparatus for single-target multi-image joint search according to an embodiment of the present invention.
Fig. 4 is a block diagram illustrating a second embodiment of an apparatus for single-target multi-image joint search according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention will be described in further detail with reference to the accompanying drawings in conjunction with the following detailed description. It should be understood that the description is intended to be exemplary only, and is not intended to limit the scope of the present invention. Moreover, in the following description, descriptions of well-known structures and techniques are omitted so as to not unnecessarily obscure the concepts of the present invention.
Please refer to fig. 1, which is a flowchart illustrating a method for joint retrieval of single-target multiple images according to a first embodiment of the present invention. The method in the embodiment is mainly used for image retrieval in various videos, particularly monitoring videos. As shown, the method includes:
step S101: a plurality of images including the same retrieval target are received.
For the retrieval target, the features extracted from a single image are not sufficient to fully describe it. When the environment or local features where the retrieval target is located change, the retrieval may be incomplete or inaccurate. In the scheme, a plurality of images of the same retrieval target are received, the characteristics of the retrieval target are comprehensively extracted from the plurality of images, and the retrieval target is subjected to multi-directional retrieval on the basis of multi-directional characteristic extraction, so that the comprehensiveness of the retrieval is improved.
Step S102: and respectively generating the difference weights of the plurality of images.
The difference (or similarity) between the images is realized by an image retrieval method, which is equivalent to performing image retrieval in a small range among all samples carried in multiple images.
Step S103: and fusing the features of the retrieval targets in the plurality of images according to the difference weights to generate a feature set.
And performing feature extraction and feature fusion on a plurality of samples of a single target to generate features which more fully describe the target.
Step S104: and searching in an image database based on the feature set to obtain a first search result.
Searching an image related to a retrieval target in the whole image database by using the fused feature set, wherein the retrieval mode is the same as that in the step S101, and the image retrieval is realized in various ways in the prior art, which is not further described herein.
Step S105: and performing weighted fusion on the first retrieval result according to the difference weight, and outputting a weighted and fused second retrieval result.
After the first retrieval result is weighted and fused by the difference weight, the accuracy of the retrieval result under multiple characteristic angles can be judged, and a more comprehensive judgment result is obtained.
In summary, the difference weights and the feature sets are extracted from a plurality of samples of the retrieval target, retrieval is performed based on the feature sets, and the difference weights are weighted into the retrieval result; and a plurality of samples of a single target are fused for retrieval, so that the retrieval result is more visual, and the precision ratio and the recall ratio are higher.
Please refer to fig. 2, which is a flowchart illustrating a method of a single-target multi-image joint search according to a second embodiment of the present invention, as shown in the figure, the method includes:
step S201: a plurality of images including the same retrieval target are received.
The multiple images are preferably images with certain differences, such as images acquired under different angles, different illumination and different backgrounds of the vehicle.
Step S202: and extracting the local features of the retrieval target, and carrying out compression coding on the local features to generate a target model.
The local features extracted from the input multiple target region images include, but are not limited to, SIFT, SURF, and the like. And performing coding compression on the extracted local features by using an offline trained model by using methods including but not limited to visual word Bag (BOW), Hamming embedding, local sensitive hashing and CDVS (compact description for image retrieval), and generating a target model.
Step S203: retrieving the target models and returning the similarity of any two target models in a sequence from high to low; the similarity is the similarity truncated by a preset similarity threshold.
During retrieval, visual word Bag (BOW), Hamming embedding, local sensitive hashing and CDVS (compact description for image retrieval) methods are used for retrieval, and the image similarity of the specified similarity threshold truncation is returned and ranked from high to low.
Step S204: and calculating the average similarity of each target model and other target models in the similarity ranking.
The average similarity
Figure GDA0002291051980000051
And i ≠ j.
Step S205: and calculating an average difference according to the average similarity, and performing L2 normalization on the average difference to obtain a difference weight.
The average degree of difference ti=1-ri,0≤i<k;
The L2 normalization specifically includes:
Figure GDA0002291051980000052
wi=ti/sum,0≤i<k;
where k denotes the total number of samples of the retrieval target, si,jRepresenting the similarity of the sample i and the sample j; w is aiRepresenting the dissimilarity weight.
The method comprises the steps of describing differences among a plurality of input samples of the same retrieval target, calculating the similarity of each sample with other samples by using an image retrieval method, wherein the similarity with other samples is high, the difference weight is small, the similarity with other samples is low, and the difference weight is great.
Suppose k (k) is given for a single target>1) Calculating the similarity of the samples with other samples in the k sample image sets by using an image retrieval algorithm to obtain the similarity s between every two samples except the k sample image setsi,jI is more than or equal to 0, j is less than k and i is not equal to j, and the average similarity r can be obtained by calculating the similarity with other samplesi
Step S206: and obtaining feature word bag centers of all features of the retrieval target through off-line training.
The off-line training mode comprises kmeans and GMM.
Step S207: and respectively calculating the membership states of the feature and the feature bag center.
And respectively calculating nearest neighbors of Euclidean distances between the features and the feature word bag center, and counting the membership condition according to the nearest neighbors.
Step S208: and removing redundant information at the center of the feature word bag according to the distance between the features to obtain a feature set.
The distance between the features is used for describing the similarity between the features, and the larger the distance is, the smaller the similarity is; the smaller the distance, the greater the similarity; if the distance is 0, it indicates that the two are identical. And calculating the distance between every two characteristics belonging to the same characteristic bag center, performing mean value fusion on all the characteristics with the distance smaller than a specified threshold value, and replacing the original characteristics with the fused characteristics.
The calculation formula of the characteristic bag-of-words center of the characteristic is as follows:
Figure GDA0002291051980000061
the calculation formula of the distance between the features is as follows:
Figure GDA0002291051980000062
the calculation formula of the mean value fusion of the features is as follows:
Figure GDA0002291051980000071
wherein f isi,j(0. ltoreq. i < n, 0. ltoreq. j < d) represents a characteristic, ci,j(i is more than or equal to 0 and less than or equal to m, and j is more than or equal to 0 and less than d) represents the center of the word bag; k represents the total number of samples of the retrieval target; n represents the total number of features, and m represents the number of feature bag centers; d represents the dimension of the feature; n isthd,iRepresenting the number of features whose distance is less than a specified threshold.
Step S209: and searching in an image database based on the feature set to obtain a first search result.
Step S210: and performing weighted fusion on the first retrieval result according to the difference weight, and outputting the sequence of the similarity from high to low after weighted fusion.
The calculation of the similarity after the weighted fusion specifically comprises the following steps:
Figure GDA0002291051980000072
wherein, wiRepresenting a difference weight, ti,jRepresenting the similarity between the image in the first retrieval result and the retrieval target, wherein k represents the total number of samples of the retrieval target; l denotes the total number of images in the first search result.
In summary, the difference weights and the feature sets are extracted from a plurality of samples of the retrieval target, retrieval is performed based on the feature sets, and the difference weights are weighted into the retrieval result; and a plurality of samples of a single target are fused for retrieval, so that the retrieval result is more visual, and the precision ratio and the recall ratio are higher. Meanwhile, the further limitation of data processing at each level is realized, the operation efficiency of the scheme is improved, and the precision ratio and the recall ratio of the picture retrieval are improved.
The following is an embodiment of a single-target multi-image joint retrieval apparatus provided in the detailed description of the present invention, and the embodiment of the single-target multi-image joint retrieval apparatus is implemented based on the above-mentioned embodiment of the single-target multi-image joint retrieval method, and please refer to the above-mentioned embodiment of the single-target multi-image joint retrieval method, which is not described in the embodiments of the single-target multi-image joint retrieval apparatus.
Referring to fig. 3, it is a block diagram illustrating a first embodiment of an apparatus for single-target multi-image joint search according to an embodiment of the present invention, as shown in the drawing, the apparatus includes:
a retrieval target receiving unit 310 configured to receive a plurality of images including the same retrieval target;
a difference weight calculating unit 320, configured to generate difference weights for the multiple images respectively;
a feature set generating unit 330, configured to fuse features of the search targets in the multiple images according to the difference weights to generate a feature set;
a primary retrieving unit 340, configured to retrieve, based on the feature set, in an image database to obtain a first retrieval result;
and a weighted fusion unit 350, configured to perform weighted fusion on the first search result according to the difference weight, and output a weighted-fused second search result.
In summary, in the cooperative work of the units, the difference weights and the feature sets are extracted from the multiple samples of the retrieval target, retrieval is performed based on the feature sets, and the difference weights are weighted into the retrieval result; and a plurality of samples of a single target are fused for retrieval, so that the retrieval result is more visual, and the precision ratio and the recall ratio are higher.
Referring to fig. 4, it is a block diagram illustrating a second embodiment of an apparatus for single-target multi-image joint search according to an embodiment of the present invention, as shown in the drawing, the apparatus includes:
a retrieval target receiving unit 310 configured to receive a plurality of images including the same retrieval target;
a difference weight calculating unit 320, configured to generate difference weights for the multiple images respectively;
a feature set generating unit 330, configured to fuse features of the search targets in the multiple images according to the difference weights to generate a feature set;
a primary retrieving unit 340, configured to retrieve, based on the feature set, in an image database to obtain a first retrieval result;
and a weighted fusion unit 350, configured to perform weighted fusion on the first search result according to the difference weight, and output a weighted-fused second search result.
Wherein the difference weight calculating unit 320 includes:
a target model generating module 321, configured to extract local features of the search target, perform compression coding on the local features, and generate a target model;
a similarity truncation module 322, configured to retrieve the target models and return a ranking of similarity of any two target models from high to low; the similarity is the similarity cut off by a preset similarity threshold;
a similarity accumulation module 323, configured to calculate an average similarity between each target model and other target models in the similarity ranking;
a difference normalization module 324, configured to calculate an average difference according to the average similarity, and perform L2 normalization on the average difference to obtain a difference weight.
Wherein the feature set generating unit 330 includes:
the offline training module 331 is configured to perform offline training to obtain feature word bag centers of all features of the search target;
a membership state judging module 332, configured to calculate membership states of the feature and the center of the feature word bag, respectively;
and the redundant eliminating module 333 is used for eliminating the redundant information at the center of the feature bag according to the distance between the features to obtain a feature set.
The weighted fusion unit 350 is specifically configured to:
and performing weighted fusion on the first retrieval result by using the difference weight, and outputting the sequence of the similarity from high to low after weighted fusion.
Wherein the average similarity
Figure GDA0002291051980000101
And i is not equal to j;
the average degree of difference ti=1-ri,0≤i<k;
The L2 normalization specifically includes:
Figure GDA0002291051980000102
wi=ti/sum,0≤i<k;
where k denotes the total number of samples of the retrieval target, si,jRepresenting the similarity of the sample i and the sample j; w is aiRepresenting the dissimilarity weight.
The off-line training mode comprises kmeans and GMM;
the membership status determining module 332 is specifically configured to:
respectively calculating nearest neighbors of Euclidean distances between the features and the feature word bag center, and counting membership conditions according to the nearest neighbors;
the redundant elimination module 333 is specifically configured to:
and calculating the distance between every two characteristics belonging to the same characteristic bag center, performing mean value fusion on all the characteristics with the distance smaller than a specified threshold value, and replacing the original characteristics with the fused characteristics.
Wherein, the calculation formula of the feature bag center of the feature is as follows:
Figure GDA0002291051980000103
the calculation formula of the distance between the features is as follows:
Figure GDA0002291051980000111
the calculation formula of the mean value fusion of the features is as follows:
Figure GDA0002291051980000112
wherein f isi,j(0. ltoreq. i < n, 0. ltoreq. j < d) represents a characteristic, ci,j(i is more than or equal to 0 and less than or equal to m, and j is more than or equal to 0 and less than d) represents the center of the word bag; k represents the total number of samples of the retrieval target; n represents the total number of features, and m represents the number of feature bag centers; d represents the dimension of the feature; n isthd,iRepresenting the number of features whose distance is less than a specified threshold.
Wherein, the calculation of the similarity after the weighted fusion specifically comprises the following steps:
Figure GDA0002291051980000113
wherein, wiRepresenting a difference weight, ti,jRepresenting the similarity between the image in the first retrieval result and the retrieval target, wherein k represents the total number of samples of the retrieval target; l denotes the total number of images in the first search result.
In summary, the above functional modules cooperate with each other to extract difference weights and feature sets from a plurality of samples of a retrieval target, perform retrieval based on the feature sets, and weight the difference weights into a retrieval result; and a plurality of samples of a single target are fused for retrieval, so that the retrieval result is more visual, and the precision ratio and the recall ratio are higher. Meanwhile, the further limitation of data processing at each level is realized, the operation efficiency of the scheme is improved, and the precision ratio and the recall ratio of the picture retrieval are improved.
It is to be understood that the above-described embodiments of the present invention are merely illustrative of or explaining the principles of the invention and are not to be construed as limiting the invention. Therefore, any modification, equivalent replacement, improvement and the like made without departing from the spirit and scope of the present invention should be included in the protection scope of the present invention. Further, it is intended that the appended claims cover all such variations and modifications as fall within the scope and boundaries of the appended claims or the equivalents of such scope and boundaries.
Although the embodiments of the present invention have been described in detail, it should be understood that various changes, substitutions, and alterations can be made hereto without departing from the spirit and scope of the invention.

Claims (14)

1. A method for single-target multi-image joint retrieval is characterized by comprising the following steps:
receiving a plurality of images comprising the same retrieval target;
respectively generating differential weights of the images;
fusing the features of the retrieval targets in the multiple images according to the difference weights to generate a feature set;
retrieving in an image database based on the feature set to obtain a first retrieval result;
and performing weighted fusion on the first retrieval result according to the difference weight, and outputting a weighted and fused second retrieval result.
2. The method of claim 1, wherein the separately generating the differential weights for the plurality of images comprises:
extracting local features of the retrieval target, and carrying out compression coding on the local features to generate a target model;
retrieving the target models and returning the similarity of any two target models in a sequence from high to low; the similarity is the similarity cut off by a preset similarity threshold;
calculating the average similarity of each target model and other target models in the similarity sequence;
and calculating an average difference according to the average similarity, and performing L2 normalization on the average difference to obtain a difference weight.
3. The method according to claim 1, wherein the fusing the features of the retrieval target in the plurality of images according to the difference weights to generate a feature set comprises:
off-line training to obtain the bag centers of the feature words of all the features of the retrieval target;
respectively calculating membership states of the characteristic and the characteristic word bag center;
and removing redundant information at the center of the feature word bag according to the distance between the features to obtain a feature set.
4. The method according to claim 1, wherein the performing weighted fusion on the first search result according to the difference weight and outputting a weighted-fused second search result comprises:
and performing weighted fusion on the first retrieval result by using the difference weight, and outputting the sequence of the similarity from high to low after weighted fusion.
5. The method of claim 2, wherein the average similarity measure
Figure FDA0002291051970000021
I is more than or equal to 0, j is less than k, and i is not equal to j;
the average degree of difference ti=1-ri,0≤i<k;
The L2 normalization specifically includes:
Figure FDA0002291051970000022
wi=ti/sum,0≤i<k;
where k denotes the total number of samples of the retrieval target, si,jRepresenting the similarity of the sample i and the sample j; w is aiRepresenting the dissimilarity weight.
6. The method of claim 3, wherein the offline training modes include kmeans and GMMs;
the respectively calculating the membership states of the feature and the feature bag center comprises the following steps:
respectively calculating nearest neighbors of Euclidean distances between the features and the feature word bag center, and counting membership conditions according to the nearest neighbors;
the removing of the redundant information of the center of the feature word bag according to the distance between the features to obtain a feature set comprises the following steps:
calculating the distance between every two characteristics belonging to the same characteristic bag center, performing mean value fusion on all characteristics with the distance smaller than a specified threshold value, and replacing the original characteristics with the fused characteristics;
wherein, the calculation formula of the feature bag center of the feature is as follows:
Figure FDA0002291051970000031
the calculation formula of the distance between the features is as follows:
Figure FDA0002291051970000032
the calculation formula of the mean value fusion of the features is as follows:
Figure FDA0002291051970000033
wherein f isi,j(0. ltoreq. i < n, 0. ltoreq. j < d) represents a characteristic, ci,j(i is more than or equal to 0 and less than or equal to m, and j is more than or equal to 0 and less than d) represents the center of the word bag; k represents the total number of samples of the retrieval target; n represents the total number of features, m represents the bag of feature wordsThe number of centers; d represents the dimension of the feature; n isthd,iRepresenting the number of features whose distance is less than a specified threshold.
7. The method according to claim 1, wherein the calculating of the weighted fused similarity comprises:
Figure FDA0002291051970000034
wherein, wiRepresenting a difference weight, ti,jRepresenting the similarity between the image in the first retrieval result and the retrieval target, wherein k represents the total number of samples of the retrieval target; l denotes the total number of images in the first search result.
8. An apparatus for single-target multi-image joint retrieval, comprising:
a retrieval target receiving unit for receiving a plurality of images including the same retrieval target;
the difference weight calculation unit is used for respectively generating difference weights of the images;
a feature set generating unit configured to fuse features of the search targets in the plurality of images according to the difference weights to generate a feature set;
the primary retrieval unit is used for retrieving in an image database based on the feature set to obtain a first retrieval result;
and the weighted fusion unit is used for performing weighted fusion on the first retrieval result according to the difference weight and outputting a weighted-fused second retrieval result.
9. The apparatus of claim 8, wherein the differential weight calculation unit comprises:
the target model generation module is used for extracting the local features of the retrieval target, carrying out compression coding on the local features and generating a target model;
the similarity truncation module is used for retrieving the target models and returning the sequence of the similarity of any two target models from high to low; the similarity is the similarity cut off by a preset similarity threshold;
the similarity accumulation module is used for calculating the average similarity of each target model and other target models in the similarity sequence;
and the difference degree normalization module is used for calculating the average difference degree according to the average similarity degree and carrying out L2 normalization on the average difference degree to obtain the difference weight.
10. The apparatus of claim 8, wherein the feature set generating unit comprises:
the off-line training module is used for off-line training to obtain the bag centers of the feature words of all the features of the retrieval target;
the membership state judging module is used for respectively calculating membership states of the characteristic and the bag center of the characteristic word;
and the redundant eliminating module is used for eliminating the redundant information at the center of the characteristic word bag according to the distance between the characteristics to obtain a characteristic set.
11. The apparatus according to claim 9, wherein the weighted fusion unit is specifically configured to:
and performing weighted fusion on the first retrieval result by using the difference weight, and outputting the sequence of the similarity from high to low after weighted fusion.
12. The apparatus of claim 9, wherein the average similarity measure
Figure FDA0002291051970000051
I is more than or equal to 0, j is less than k, and i is not equal to j;
the average degree of difference ti=1-ri,0≤i<k;
The L2 normalization specifically includes:
Figure FDA0002291051970000052
wi=ti/sum,0≤i<k;
where k denotes the total number of samples of the retrieval target, si,jRepresenting the similarity of the sample i and the sample j; w is aiRepresenting the dissimilarity weight.
13. The apparatus of claim 10, wherein the offline training modes comprise kmeans and GMMs;
the membership state judgment module is specifically configured to:
respectively calculating nearest neighbors of Euclidean distances between the features and the feature word bag center, and counting membership conditions according to the nearest neighbors;
the redundancy eliminating module is specifically used for:
calculating the distance between every two characteristics belonging to the same characteristic bag center, performing mean value fusion on all characteristics with the distance smaller than a specified threshold value, and replacing the original characteristics with the fused characteristics;
wherein, the calculation formula of the feature bag center of the feature is as follows:
Figure FDA0002291051970000061
the calculation formula of the distance between the features is as follows:
Figure FDA0002291051970000062
the calculation formula of the mean value fusion of the features is as follows:
Figure FDA0002291051970000063
wherein f isi,j(0. ltoreq. i < n, 0. ltoreq. j < d) represents a characteristic, ci,j(i is more than or equal to 0 and less than or equal to m, and j is more than or equal to 0 and less than d) represents the center of the word bag; k represents the total number of samples of the retrieval target; n represents TeThe total number of the characters, m represents the number of the centers of the character word bags; d represents the dimension of the feature; n isthd,iRepresenting the number of features whose distance is less than a specified threshold.
14. The apparatus according to claim 8, wherein the calculation of the similarity after weighted fusion specifically comprises:
Figure FDA0002291051970000071
wherein, wiRepresenting a difference weight, ti,jRepresenting the similarity between the image in the first retrieval result and the retrieval target, wherein k represents the total number of samples of the retrieval target; l denotes the total number of images in the first search result.
CN201510672504.2A 2015-09-24 2015-10-16 Single-target multi-image joint retrieval method and device Active CN106557533B (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN2015106167309 2015-09-24
CN201510616730 2015-09-24

Publications (2)

Publication Number Publication Date
CN106557533A CN106557533A (en) 2017-04-05
CN106557533B true CN106557533B (en) 2020-03-06

Family

ID=58417898

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510672504.2A Active CN106557533B (en) 2015-09-24 2015-10-16 Single-target multi-image joint retrieval method and device

Country Status (1)

Country Link
CN (1) CN106557533B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110083732B (en) * 2019-03-12 2021-08-31 浙江大华技术股份有限公司 Picture retrieval method and device and computer storage medium
CN111815284B (en) * 2020-07-14 2022-10-21 广东冶建施工图审查中心有限公司 Engineering flow building system based on data analysis

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103793721A (en) * 2014-03-04 2014-05-14 武汉大学 Pedestrian repeat recognition method and system based on area related feedback
CN103810252A (en) * 2014-01-21 2014-05-21 南京信息工程大学 Image retrieval method based on group sparse feature selection
CN104281572A (en) * 2013-07-01 2015-01-14 中国科学院计算技术研究所 Target matching method and system based on mutual information
CN104298758A (en) * 2014-10-22 2015-01-21 天津大学 Multi-perspective target retrieval method

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2897059A4 (en) * 2012-09-13 2016-07-06 Ntt Docomo Inc User interface device, search method, and program

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104281572A (en) * 2013-07-01 2015-01-14 中国科学院计算技术研究所 Target matching method and system based on mutual information
CN103810252A (en) * 2014-01-21 2014-05-21 南京信息工程大学 Image retrieval method based on group sparse feature selection
CN103793721A (en) * 2014-03-04 2014-05-14 武汉大学 Pedestrian repeat recognition method and system based on area related feedback
CN104298758A (en) * 2014-10-22 2015-01-21 天津大学 Multi-perspective target retrieval method

Also Published As

Publication number Publication date
CN106557533A (en) 2017-04-05

Similar Documents

Publication Publication Date Title
CN109815364B (en) Method and system for extracting, storing and retrieving mass video features
Li et al. Contextual bag-of-words for visual categorization
CN104035949B (en) Similarity data retrieval method based on locality sensitive hashing (LASH) improved algorithm
Zheng et al. $\mathcal {L} _p $-Norm IDF for Scalable Image Retrieval
Zhou et al. Towards codebook-free: Scalable cascaded hashing for mobile image search
CN102254015B (en) Image retrieval method based on visual phrases
US20140193077A1 (en) Image retrieval apparatus, image retrieval method, query image providing apparatus, query image providing method, and program
CN109635686B (en) Two-stage pedestrian searching method combining human face and appearance
CN108596010B (en) Implementation method of pedestrian re-identification system
KR20130142191A (en) Robust feature matching for visual search
CN104794219A (en) Scene retrieval method based on geographical position information
CN111507350B (en) Text recognition method and device
CN110503643B (en) Target detection method and device based on multi-scale rapid scene retrieval
CN111460961A (en) CDVS-based similarity graph clustering static video summarization method
CN110046568B (en) Video action recognition method based on time perception structure
CN115062186B (en) Video content retrieval method, device, equipment and storage medium
TW202109312A (en) Image feature extraction method, network training method, electronic device and computer readable storage medium
CN109241315B (en) Rapid face retrieval method based on deep learning
CN106557533B (en) Single-target multi-image joint retrieval method and device
CN104778272B (en) A kind of picture position method of estimation excavated based on region with space encoding
CN108764258B (en) Optimal image set selection method for group image insertion
CN113723558A (en) Remote sensing image small sample ship detection method based on attention mechanism
CN109918529A (en) A kind of image search method based on the quantization of tree-like cluster vectors
CN113705310A (en) Feature learning method, target object identification method and corresponding device
CN104699783A (en) Social image searching method allowing adaptive adjustment and based on personalized vision dictionary

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant