CN110059206A - A kind of extensive hashing image search method based on depth representative learning - Google Patents

A kind of extensive hashing image search method based on depth representative learning Download PDF

Info

Publication number
CN110059206A
CN110059206A CN201910249642.8A CN201910249642A CN110059206A CN 110059206 A CN110059206 A CN 110059206A CN 201910249642 A CN201910249642 A CN 201910249642A CN 110059206 A CN110059206 A CN 110059206A
Authority
CN
China
Prior art keywords
loss function
label
layer
data
hash
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910249642.8A
Other languages
Chinese (zh)
Inventor
王祥丰
胡慷
孔桦桦
田伟
陈寅峰
李鑫
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Enjoyor Co Ltd
Original Assignee
Enjoyor Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Enjoyor Co Ltd filed Critical Enjoyor Co Ltd
Priority to CN201910249642.8A priority Critical patent/CN110059206A/en
Publication of CN110059206A publication Critical patent/CN110059206A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/53Querying
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Library & Information Science (AREA)
  • Image Analysis (AREA)

Abstract

A kind of extensive hashing image search method based on depth representative learning, the post-processing of Optimization Learning and search result including data prediction, based on semi-supervised convolutional neural networks layer building, based on pairs of Hash loss function design step loss function.The present invention carries out the extraction of characteristics of image and the study of hash function using depth convolutional neural networks layer and full articulamentum, devise the assembling loss function for intersecting entropy loss item, triple loss item and pseudo label loss item and constituting of tape label, it is optimized using the stochastic gradient descent method with momentum, computational efficiency with higher finally realizes the unified image retrieval performance of accuracy and speed.The present invention carries out picture search problem to carry out High Efficiency Modeling under the basis based on image data structure and label under the status of picture search field fast development, the model accuracy and more optimized inquiry velocity effectively improved.

Description

A kind of extensive hashing image search method based on depth representative learning
Technical field
The invention belongs to field of image search, are related to a kind of large-scale image search method.
Background technique
With the rapid development of the technologies such as big data technology, development of Mobile Internet technology, technology of Internet of things, image, video etc. Acquisition, convergence and the storage of multimedia resource are more and more convenient.In practical applications, 80% or so data be all with document, The unstructured data of the forms such as image, video, audio storage, and same such unstructured data all presses index every year Increase about 3/5ths.In unstructured data, image be again it is most important, occupy enormous proportions, but at the same time Important information is contained.Computer vision is primarily upon the research of image data, therefore also naturally enough becomes machine learning One of with the most popular research point of artificial intelligence technology.
Application problem in face of various complexity and the amount of images growth rate that is exceedingly fast promote traditional images analysis to handle Technology encounters very big challenge, especially huge resource, superelevation dimension, mass memory, slowly in terms of challenge. To solve the problems, such as this some column, a variety of approximate algorithms with low time complexity are suggested, wherein to be based on hash algorithm The algorithm of representative obtains a large amount of concerns, because it has many advantages, such as constant enquiry time and a small amount of compression binary code storage.But A large amount of image data is all not no label, and especially in fields such as picture search, the image data of only only a few is had Label information, this is also to consume what a large amount of manpower and material resources obtained.How to merge has flag data and a large amount of nothings using a small amount of The image search method of flag data is the method with Great significance and actual demand.Design targetedly efficient needle Derivation algorithm to large-scale image search is vital.
Hash algorithm is typical nearest neighbor search method, is mainly used for solving memory space recited above greatly and retrieves The problem of time length.In hash algorithm, common target is one mapping function W of study, and sample is expressed as a string of fixations The binary-coding of length, encoding usually using {+1, -1 } or {+1,0 } indicates bit each, so that semantically similar Sample can obtain similar coding, we are usually using Hamming distance from the similitude between measurement binary-coding.
In initial work, local sensitivity hash algorithm (LSH) is a kind of popular algorithm, the key of LSH algorithm Part is to divide space using the hyperplane being randomly generated, and it is every to determine to fall in the position of hyperplane according to sample One value, LSH require similar sample to fall among the same Hash bucket with greater probability value, without similar sample with Minimum probability value is fallen among the same Hash bucket, and by using bit multiple, final LSH can reach satisfactory inspection Suo Xiaoguo, LSH method have stringent theoretical proof to guarantee its performance.But in order to obtain shorter more compact coding And better retrieval performance, researcher consider the label information of image from different objective function is constructed, and choose different excellent Change method must further improve the performance of Hash retrieval using linearly or nonlinearly model etc..
The appearance of deep neural network causes huge repercussion in computer vision field, and researcher is unnecessary to go again The complicated inefficient expression feature of engineer, but neural network oneself is allowed to go to learn the characterization of image.The present invention is exactly to combine Deep neural network designs a kind of semi-supervised Hash searching algorithm, the Hash searching algorithm of the remote ultra-traditional of retrieval rate.
Summary of the invention
In order to overcome the lower deficiency of accuracy rate of existing large-scale image search method, the present invention provides a kind of accurate The higher extensive hashing image search method based on depth representative learning of rate.
The technical solution adopted by the present invention to solve the technical problems is:
A kind of extensive hashing image search method based on depth representative learning, comprising the following steps:
Step 1: data prediction;
Step 2: deep neural network constructs, using pretreated labeling in data prediction step or without label Data carry out deep neural network model building;
Step 3: Hash loss function designs, the Hash loss function is by softmax cross entropy loss function, ternary Group loss function, pseudo label loss function three parts composition.
Step 4: deep neural network training and model optimization derive letter for the Hash loss function of step 3 Several gradients carries out the update of parameter until convergence to entire depth neural network model using the gradient descent method with momentum;
Step 5, the deep neural network model after input picture to training, is calculated binary system Hash codes to data Image in library calculates Hamming distances, returns to candidate result of the most like image as picture search.
Preferably, the step 5 further includes, the candidate result is the result of coarseness;Select layer mind second from the bottom Output through network is further calculated the Euclidean distance of these characterizations in candidate result, is arranged again with this as characterization Sequence, the final result are exactly fine-grained search result.
Preferably, the data prediction step includes but is not limited to the standardization of image normalizing, drop in the step 1 It makes an uproar, deblurring, image data augmentation preliminary treatment, further includes whether having class declaration according to data to be divided into tape label data With without label data, for there is the data of label, it is also necessary to construct triple structure.
Preferably, the output of neural network the last layer includes two parts in the step 3: Hash codes and classification Value;The output of neural network the last layer includes two parts: Hash codes and class label;The class label intersects as softmax The input of entropy loss function is converted into the class probability between [0,1] by softmax function, then passes through cross entropy loss function Obtain penalty values to the end;The Hash codes are the inputs of triple loss function and pseudo label loss function.
The softmax cross entropy loss function:
JS(xi)=- yilogP(xi)
The wherein index of i representative sample, x represent data sample, and y represents label, P (xi) represent i-th of sample and belong to y's Probability value;
The triple loss function:
Wherein, i is sample index, and x represents data sample,Indicate xLSample with same label,Represent and xLSample with different labels, h (*) represent the corresponding output of network, mtFor threshold parameter;
The pseudo label loss function:
Wherein, i, j are similarly sample index, for the data of not tape label, assign the maximum label of probability value to its work For pseudo label, mpFor threshold parameter, the maximum distance of two different classes of samples is represented, A (i, j)=1 indicates that sample i, j belong to Same class, A (i, j)=0 indicate that sample i, j belong to inhomogeneity, and h (*) represents the corresponding output of network.The Hash loss function Are as follows:
Wherein JS(xi) it is softmax cross entropy loss function,It is triple loss function,It is pseudo label loss function, p is pseudo label, and α, β and γ are weight parameters.
Preferably, for softmax cross entropy loss function, passing through deep learning lower portion in the step 4 Automatic derivation;For triple loss function and pseudo label loss function, derivation is re-started using back-propagation algorithm.
Preferably, in an iterative process, being trained optimization to tape label data first, first not in the step 4 Using pseudo label loss function;When loss drops to given threshold, unlabeled exemplars data are added, are lost in conjunction with pseudo label Function is iterated optimization to Hash loss function until convergence.
Preferably, the deep neural network uses basic neural network framework in the step 2, pass through stacking Convolutional layer, pond layer and full articulamentum form core network.In the neural network framework, 2 layers are added to connect entirely using 5 layers of convolutional layer The mode of layer is connect, and distinguishes Hash coding layer and classification layer simultaneously in the last layer;Network parameter is set as, first layer convolution Layer: convolution kernel size is 11x11, stride 4, port number 64, activation primitive Relu;Pool layers of Max: core size is 3x3, stride 2;Second layer convolutional layer: convolution kernel size is 5x5, and stride 1, port number 192, activation primitive is Relu.Pool layers of Max: core size is 3x3, stride 2;Third layer convolutional layer: convolution kernel size be 3x3, stride 1, Port number is 384, activation primitive Relu;4th layer of convolutional layer: convolution kernel size is 5x5, stride 1, and port number is 256, activation primitive Relu;Layer 5 convolutional layer: convolution kernel size is 3x3, stride 1, port number 256, activation letter Number is Relu;The full articulamentum neuron nodal point number of first layer is 2048;The full articulamentum neuron nodal point number of the second layer is Hash codes The sum of digit and classification number.
Beneficial effects of the present invention are mainly manifested in:
1, in the network model building based on depth, the network architecture of similar AlexNet is selected, directly by original graph As the input as model, advantage is that neural network can acquire the characterization of better image data, has stronger extensive Performance;
2, ingenious to combine triple loss in the design of Hash loss function, softmax intersects entropy loss and pseudo label damage Item is lost, the study of model multi-angle can be conducive to, to obtain higher robustness.
3, in the optimization of model and the update of parameter, the derived function of loss function and the training side of model are given Method can reduce manually flag data in the case where guaranteeing test accuracy rate in conjunction with this semi-supervised mode of learning The required time, to save human cost.
4, in the step of image retrieval, we have proposed multi-level sort methods from thick to thin, can allow in the time In the case where, further increase the accuracy rate of retrieval.
Detailed description of the invention
Fig. 1 is the flow chart of the extensive hashing image search method the present invention is based on depth representative learning.
Fig. 2 is the network of the extensive hashing image search method the present invention is based on depth representative learning.
Specific embodiment
The invention will be further described below in conjunction with the accompanying drawings.
Referring to Figures 1 and 2, a kind of extensive hashing image search method based on depth representative learning, by image data Structure and label condition driving effectively modeling proposed a kind of based on depthmeter on the basis of existing picture search algorithm The extensive hashing image search method of study is levied, its advantages are that of obtaining high accuracy and quick search speed and rule Mould scalability and less memory space.
The extensive hashing image search method the following steps are included:
Step 1: data prediction
In data prediction part, first have to class declaration whether is had according to the data in data set, for example, class name or Person one-hot coding, is divided into tape label data and without label data, for there is the data of label, it is also necessary to construct ternary Group structure, that is, for sample xL, choose another sample for having same labelAnother is chosen with different marks The sample of labelThe data for forming a triple structure, the input as network.It is also desirable to all data sets Standardization processing is carried out, the operation for subtracting mean value divided by standard deviation is carried out, controls the numeric distribution of sample.In the situation of sample deficiency Under, need to carry out data augmentation operation, including rotated, translated to picture, fuzzy noise etc., and the width of image is high Resize to 224.
Step 2: deep neural network constructs
Deep neural network model is carried out using pretreated labeling in data prediction step or without label data Building, herein, we still take basic neural network framework, by stacking convolutional layer, pond layer and full articulamentum group At core network.In network model building, the network structure of AlexNet is copied, adds 2 layers of full articulamentum using 5 layers of convolutional layer Mode, and the last layer simultaneously distinguish Hash coding layer and classification layer.Specific network parameter is set as, first layer convolution Layer: convolution kernel size is 11x11, stride 4, port number 64, activation primitive Relu.Pool layers of Max: core size is 3x3, stride 2.Second layer convolutional layer: convolution kernel size is 5x5, and stride 1, port number 192, activation primitive is Relu.Pool layers of Max: core size is 3x3, stride 2.Third layer convolutional layer: convolution kernel size be 3x3, stride 1, Port number is 384, activation primitive Relu.4th layer of convolutional layer: convolution kernel size is 5x5, stride 1, and port number is 256, activation primitive Relu.Layer 5 convolutional layer: convolution kernel size is 3x3, stride 1, port number 256, activation letter Number is Relu.The full articulamentum neuron nodal point number of first layer is 2048.The full articulamentum neuron nodal point number of the second layer is Hash codes The sum of digit and classification number, such as Hash code bit number are 48, share 10 classes, then the number of nodes of full articulamentum is 58.Specific net Network structure chart is as shown in Figure 2.
Step 3: Hash loss function designs
In the design of loss function, it is divided into three parts, the softmax cross entropy loss function of tape label, triple Loss function, pairs of pseudo label loss function.
The output of neural network the last layer includes two parts: Hash codes and class label.
Input of the class label as softmax will be converted into the class probability between [0,1] by softmax function, then be led to It crosses cross entropy objective function and obtains penalty values to the end, softmax cross entropy loss function is most common Classification Loss function.
Hash codes part is the input of triple loss function item and pairs of pseudo label loss item, and triple loss is not It is same as another loss function that softmax intersects entropy loss, it is characterized in that identical two samples of the meaning of one's words is promoted to measure Under space from as far as possible closely, and different two samples of the meaning of one's words under metric space from must as far as possible far.It is pairs of to be Tag entry loss item can regard certain a part of triple loss item in the case where the meaning of one's words is identical or different as.
First part is that commonly used softmax intersects entropy loss in image classification, for each sample, our all mistakes Softmax obtains the probability value P of classification, the calculating target function in a manner of cross entropy.It is as follows:
JS(xi)=- yilog P(xi)
The wherein index of i representative sample, x representative sample, y represent label, P (xi) represent the probability that i-th of sample belongs to y Value;
Second part is triple loss item, for the triple data of the tape label constructed in step 1, Wo Menshe The distance that a mode measures them is counted.Each sample x can have output valve h (x), for identical exemplar xLWithThe distance between they are minimized, are calculated with Euclidean distance.Conversely, then maximizing the distance between they.It is as follows:
Wherein, i is sample index, and x represents data sample,Indicate xLSample with same label,Representative and xL Sample with different labels, h (*) represent the corresponding output of network, mtFor threshold parameter;
Part III is pseudo label loss item, and for unlabelled sample, we can obtain institute by neural network output Belong to the probability value of each classification, the pseudo label that we assign the classification where maximum value as it constructs pseudo- as second part The distance between label measurement.It is as follows:
Wherein, i, j are similarly sample index, mpFor threshold parameter, the maximum distance of two different classes of samples, A are represented (i, j)=1 indicates that sample i, j belong to same class, and A (i, j)=0 indicates that sample i, j belong to inhomogeneity, and it is corresponding that h (*) represents network Output.
Hash loss function specifically:
Wherein JS(xi) be tape label softmax cross entropy loss function,It is triple damage Function is lost,It is pseudo label loss function, p is pseudo label.Three loss functions It can be regulated and controled by weight between, wherein α, β and γ be as weight parameter, we enable α=1 herein, β=1, γ= 0.5。
Step 4: deep neural network training and model optimization
For the loss function of step 3, the gradient of derived function loss function utilizes the gradient descent method pair with momentum Entire model carries out the update of parameter until convergence.
Wherein, we are trained convergence to the sample of tape label first, that is, loss decline about to 0.2 with Under, it then is added unlabeled exemplars, alternately be trained.This semi-supervised alternating training method can be very big The time required for marker samples is reduced in degree, saves cost of human resources.The side SGD with momentum is selected in terms of optimizer Method, learning rate are set as 0.001, and momentum is set as 0.9.Common gradient declines updating method are as follows:
Wherein WtFor the parameter of t moment neural network, ▽ J is the derivative of loss function J.
Momentum gradient descent method is as follows:
Wherein θ is momentum parameter, replaces former gradient to carry out parameter update using gradient after exponent-weighted average.Because every The information of gradient before gradient after a exponent-weighted average contains.
For softmax cross entropy loss function, pass through the automatic derivation of deep learning lower portion;Triple is lost Function and pseudo label loss function, re-start derivation using back-propagation algorithm;
For triple loss function derivation:
Wherein I represents indicator function, when C is True, then for 1 on the contrary be then -1;
For pseudo label loss function derivation:
As A (i, j)=1:
As A (i, j)=0:
Optimization is iterated to objective function until convergence.
Step 5: image retrieval
After the completion of model training, we keep the Hash coding of training data.Give a test query picture, mould Type can export the binary-coding that it is predicted.The Hash of the binary-coding of prediction and training data is encoded and carries out hamming by we The measurement of distance returns to candidate result of the most like image as picture search, is waited according to apart from size given threshold Selected works prepare.
If it is required that higher inquiry precision, can carry out finer sequence, to the inverse of Candidate Set and training sample Second layer feature carries out Euclidean distance comparison, to resequence, obtains more accurately retrieval effectiveness.It is looked into for example, we give Ask image Iq, candidate pool P has been obtained after hamming sorts, we are that the higher front layer of dimension is used to connect output entirely as candidate Collect the detailed characterizations of P, we use VqAnd Vi PRespectively represent query image IqWith the feature vector of candidate pool P, we pass through Europe Formula distance measure P in i-th of sample and query image similarity S, as follows: where S is smaller, represent two samples it Between similarity it is higher, we can reorder to candidate pool by S, obtain more preferably query result.
si=| | Vq-Vi P||
Wherein, S is smaller, and the similarity represented between two samples is higher, we can reset candidate pool by S Sequence obtains more preferably query result.
The scheme of the present embodiment, using depth convolutional neural networks layer and full articulamentum carry out characteristics of image extraction and The study of hash function devises the intersection entropy loss item, triple loss item and pseudo label damage of tape label on this basis The assembling loss function that item is constituted is lost, is optimized using the stochastic gradient descent method with momentum, calculating with higher Efficiency, finally, the searching method that the present invention is combined using coarseness and fine granularity, realizes the unified image of accuracy and speed with this Retrieval performance.The present invention carries out based on image data picture search problem under the status of picture search field fast development High Efficiency Modeling, the model accuracy effectively improved and more optimized inquiry velocity are carried out under the basis of structure and label.The present invention The algorithm proposed has better accuracy and robustness than traditional several image retrieval algorithms, and the present invention can be used for Large-scale image retrieval scene, serves intelligent searching system, such as: search engine to scheme to search figure function, electric business website Similar commercial articles searching, the individualized content of social platform recommends etc..
Protection content of the invention is not limited to above embodiments.Without departing from the spirit and scope of the invention, originally Field technical staff it is conceivable that variation and advantage be all included in the present invention, and with appended claims be protect Protect range.

Claims (10)

1. a kind of extensive hashing image search method based on depth representative learning, which is characterized in that the method includes with Lower step:
Step 1: pre-processing image data;
Step 2: deep neural network constructs, using pretreated labeling in data prediction step or without label data Carry out deep neural network model building;
Step 3: Hash loss function designs, the Hash loss function is damaged by softmax cross entropy loss function, triple Lose function, pseudo label loss function three parts composition.
Step 4: deep neural network training and model optimization, for the Hash loss function of step 3, derived function loss The gradient of function carries out the update of parameter to entire depth neural network model until receiving using the gradient descent method with momentum It holds back.
Step 5, the deep neural network model after input picture to training, is calculated binary system Hash codes in database Image calculate Hamming distances, return to candidate result of the most like image as picture search.
2. a kind of extensive hashing image search method based on depth representative learning as described in claim 1, feature exist In the step 5 further includes that the candidate result is the result of coarseness;The output of layer neural network second from the bottom is selected to make For characterization, the Euclidean distance of these characterizations is further calculated in candidate result, is resequenced with this, the knot finally obtained Fruit is exactly fine-grained search result.
3. a kind of extensive hashing image search method based on depth representative learning as claimed in claim 1 or 2, feature It is, in the step 1, the data prediction step includes but is not limited to the standardization of image normalizing, noise reduction, deblurring, figure It further include whether having class declaration according to data to be divided into tape label data and without number of tags as the preliminary treatment of data augmentation According to for there is the data of label, it is also necessary to construct triple structure.
4. a kind of extensive hashing image search method based on depth representative learning as claimed in claim 3, feature exist In in the step 3, the output of neural network the last layer includes two parts: Hash codes and class label;The class label is made For the input of softmax cross entropy loss function, the class probability between [0,1] is converted by softmax function, then passes through friendship Pitch the penalty values of entropy loss function call to the end;The Hash codes are the defeated of triple loss function and pseudo label loss function Enter.
5. a kind of extensive hashing image search method based on depth representative learning as claimed in claim 4, feature exist In the softmax cross entropy loss function:
JS(xi)=- yilogP(xi)
The wherein index of i representative sample, x represent data sample, and y represents label, P (xi) represent the probability that i-th of sample belongs to y Value;
The triple loss function:
Wherein, i is sample index, and x represents data sample,Indicate xLSample with same label,Representative and xLIt has The sample of different labels, h (*) represent the corresponding output of network, mtFor threshold parameter;
The pseudo label loss function:
Wherein, i, j are similarly sample index, for the data of not tape label, assign the maximum label of probability value to it as puppet Label, mpFor threshold parameter, the maximum distance of two different classes of samples is represented, A (i, j)=1 indicates sample i, and j belongs to same Class, A (i, j)=0 indicate that sample i, j belong to inhomogeneity, and h (*) represents the corresponding output of network.
6. a kind of extensive hashing image search method based on depth representative learning as claimed in claim 5, feature exist In the Hash loss function are as follows:
Wherein JS(xi) it is softmax cross entropy loss function,It is triple loss function,It is pseudo label loss function, p is pseudo label, and α, β and γ are weight parameters.
7. a kind of extensive hashing image search method based on depth representative learning as claimed in claim 1 or 2, feature It is, in the step 4, for softmax cross entropy loss function, passes through the automatic derivation of deep learning lower portion;For Triple loss function and pseudo label loss function, re-start derivation using back-propagation algorithm.
8. a kind of extensive hashing image search method based on depth representative learning as claimed in claim 1 or 2, feature It is, in the step 4, in an iterative process, optimization is trained to tape label data first, does not first use pseudo label to damage Lose function;When loss drops to given threshold, unlabeled exemplars data are added, in conjunction with pseudo label loss function, to Hash Loss function is iterated optimization until convergence.
9. a kind of extensive hashing image search method based on depth representative learning as claimed in claim 1 or 2, feature It is, in the step 2, the deep neural network uses basic neural network framework, by stacking convolutional layer, Chi Hua Layer and full articulamentum form core network.
10. a kind of extensive hashing image search method based on depth representative learning as claimed in claim 9, feature exist In the neural network framework distinguishes Hash in such a way that 5 layers of convolutional layer add 2 layers of full articulamentum, and in the last layer simultaneously Coding layer and classification layer;Network parameter is set as, first layer convolutional layer: convolution kernel size is 11x11, stride 4, port number It is 64, activation primitive Relu;Pool layers of Max: core size is 3x3, stride 2;Second layer convolutional layer: convolution kernel size For 5x5, stride 1, port number 192, pool layers of activation primitive Relu, Max: core size is 3x3, stride 2; Third layer convolutional layer: convolution kernel size is 3x3, stride 1, port number 384, activation primitive Relu;4th layer of convolution Layer: convolution kernel size is 5x5, stride 1, port number 256, activation primitive Relu;Layer 5 convolutional layer: convolution kernel is big Small is 3x3, stride 1, port number 256, activation primitive Relu;The full articulamentum neuron nodal point number of first layer is 2048;The full articulamentum neuron nodal point number of the second layer is the sum of Hash code bit number and classification number.
CN201910249642.8A 2019-03-29 2019-03-29 A kind of extensive hashing image search method based on depth representative learning Pending CN110059206A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910249642.8A CN110059206A (en) 2019-03-29 2019-03-29 A kind of extensive hashing image search method based on depth representative learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910249642.8A CN110059206A (en) 2019-03-29 2019-03-29 A kind of extensive hashing image search method based on depth representative learning

Publications (1)

Publication Number Publication Date
CN110059206A true CN110059206A (en) 2019-07-26

Family

ID=67317886

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910249642.8A Pending CN110059206A (en) 2019-03-29 2019-03-29 A kind of extensive hashing image search method based on depth representative learning

Country Status (1)

Country Link
CN (1) CN110059206A (en)

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110751267A (en) * 2019-09-30 2020-02-04 京东城市(北京)数字科技有限公司 Neural network structure searching method, training method, device and storage medium
CN110929099A (en) * 2019-11-28 2020-03-27 杭州趣维科技有限公司 Short video frame semantic extraction method and system based on multitask learning
CN111274430A (en) * 2020-01-19 2020-06-12 易拍全球(北京)科贸有限公司 Porcelain field image retrieval algorithm based on feature reconstruction supervision
CN111274445A (en) * 2020-01-20 2020-06-12 山东建筑大学 Similar video content retrieval method and system based on triple deep learning
CN111339342A (en) * 2019-11-08 2020-06-26 深圳北航新兴产业技术研究院 Three-dimensional model retrieval method based on angle ternary center loss
CN111460200A (en) * 2020-03-04 2020-07-28 西北大学 Image retrieval method and model based on multitask deep learning and construction method thereof
CN111860620A (en) * 2020-07-02 2020-10-30 苏州富鑫林光电科技有限公司 Multilayer hierarchical neural network architecture system for deep learning
CN112000827A (en) * 2020-08-27 2020-11-27 广州搜料亿网络科技有限公司 Hardware image retrieval method and system based on deep learning
CN112329921A (en) * 2020-11-11 2021-02-05 浙江大学 Diuretic dose reasoning device based on deep characterization learning and reinforcement learning
CN112528066A (en) * 2020-12-18 2021-03-19 广东工业大学 Trademark retrieval method and system based on attention mechanism, computer equipment and storage medium
CN115062180A (en) * 2022-08-15 2022-09-16 阿里巴巴(中国)有限公司 Object query method, electronic device and storage medium
CN115292535A (en) * 2022-08-24 2022-11-04 合肥市正茂科技有限公司 Hierarchical vehicle image retrieval method based on depth polarization Hash
CN117076454A (en) * 2023-08-21 2023-11-17 广州地铁集团有限公司 Engineering quality acceptance form data structured storage method and system

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103226585A (en) * 2013-04-10 2013-07-31 大连理工大学 Self-adaptation Hash rearrangement method for image retrieval
CN105512289A (en) * 2015-12-07 2016-04-20 郑州金惠计算机***工程有限公司 Image retrieval method based on deep learning and Hash
CN106649688A (en) * 2016-12-16 2017-05-10 深圳市华尊科技股份有限公司 Image retrieval method and terminal
CN106980641A (en) * 2017-02-09 2017-07-25 上海交通大学 The quick picture retrieval system of unsupervised Hash and method based on convolutional neural networks
CN107273478A (en) * 2017-06-09 2017-10-20 华东师范大学 A kind of semi-supervised hashing image searching method based on Group Lasso
CN107992611A (en) * 2017-12-15 2018-05-04 清华大学 The high dimensional data search method and system of hash method are distributed based on Cauchy
CN108764281A (en) * 2018-04-18 2018-11-06 华南理工大学 A kind of image classification method learning across task depth network based on semi-supervised step certainly
CN109165306A (en) * 2018-08-09 2019-01-08 长沙理工大学 Image search method based on the study of multitask Hash
CN109871461A (en) * 2019-02-13 2019-06-11 华南理工大学 The large-scale image sub-block search method to be reordered based on depth Hash network and sub-block

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103226585A (en) * 2013-04-10 2013-07-31 大连理工大学 Self-adaptation Hash rearrangement method for image retrieval
CN105512289A (en) * 2015-12-07 2016-04-20 郑州金惠计算机***工程有限公司 Image retrieval method based on deep learning and Hash
CN106649688A (en) * 2016-12-16 2017-05-10 深圳市华尊科技股份有限公司 Image retrieval method and terminal
CN106980641A (en) * 2017-02-09 2017-07-25 上海交通大学 The quick picture retrieval system of unsupervised Hash and method based on convolutional neural networks
CN107273478A (en) * 2017-06-09 2017-10-20 华东师范大学 A kind of semi-supervised hashing image searching method based on Group Lasso
CN107992611A (en) * 2017-12-15 2018-05-04 清华大学 The high dimensional data search method and system of hash method are distributed based on Cauchy
CN108764281A (en) * 2018-04-18 2018-11-06 华南理工大学 A kind of image classification method learning across task depth network based on semi-supervised step certainly
CN109165306A (en) * 2018-08-09 2019-01-08 长沙理工大学 Image search method based on the study of multitask Hash
CN109871461A (en) * 2019-02-13 2019-06-11 华南理工大学 The large-scale image sub-block search method to be reordered based on depth Hash network and sub-block

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
JIAN ZHANG 等: "SSDH: Semi-supervised Deep Hashing for Large Scale Image Retrieval", 《IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY》 *
YUEFU ZHOU 等: "Deep hashing with triplet quantization loss", 《2017 IEEE VISUAL COMMUNICATIONS AND IMAGE PROCESSING (VCIP)》 *
彭天强 等: "基于深度卷积神经网络和二进制哈希学习的图像检索方法", 《电子与信息学报》 *
闭蓉: "基于深度哈希的图片检索***的设计与实现", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110751267B (en) * 2019-09-30 2021-03-30 京东城市(北京)数字科技有限公司 Neural network structure searching method, training method, device and storage medium
CN110751267A (en) * 2019-09-30 2020-02-04 京东城市(北京)数字科技有限公司 Neural network structure searching method, training method, device and storage medium
CN111339342A (en) * 2019-11-08 2020-06-26 深圳北航新兴产业技术研究院 Three-dimensional model retrieval method based on angle ternary center loss
CN111339342B (en) * 2019-11-08 2023-05-05 深圳北航新兴产业技术研究院 Three-dimensional model retrieval method based on angle ternary center loss
CN110929099A (en) * 2019-11-28 2020-03-27 杭州趣维科技有限公司 Short video frame semantic extraction method and system based on multitask learning
CN111274430A (en) * 2020-01-19 2020-06-12 易拍全球(北京)科贸有限公司 Porcelain field image retrieval algorithm based on feature reconstruction supervision
CN111274445A (en) * 2020-01-20 2020-06-12 山东建筑大学 Similar video content retrieval method and system based on triple deep learning
CN111460200A (en) * 2020-03-04 2020-07-28 西北大学 Image retrieval method and model based on multitask deep learning and construction method thereof
CN111860620A (en) * 2020-07-02 2020-10-30 苏州富鑫林光电科技有限公司 Multilayer hierarchical neural network architecture system for deep learning
CN112000827A (en) * 2020-08-27 2020-11-27 广州搜料亿网络科技有限公司 Hardware image retrieval method and system based on deep learning
CN112329921A (en) * 2020-11-11 2021-02-05 浙江大学 Diuretic dose reasoning device based on deep characterization learning and reinforcement learning
CN112329921B (en) * 2020-11-11 2023-11-14 浙江大学 Diuretic dose reasoning equipment based on deep characterization learning and reinforcement learning
CN112528066A (en) * 2020-12-18 2021-03-19 广东工业大学 Trademark retrieval method and system based on attention mechanism, computer equipment and storage medium
CN112528066B (en) * 2020-12-18 2023-08-04 广东工业大学 Trademark retrieval method, system, computer device and storage medium based on attention mechanism
CN115062180A (en) * 2022-08-15 2022-09-16 阿里巴巴(中国)有限公司 Object query method, electronic device and storage medium
CN115292535A (en) * 2022-08-24 2022-11-04 合肥市正茂科技有限公司 Hierarchical vehicle image retrieval method based on depth polarization Hash
CN117076454A (en) * 2023-08-21 2023-11-17 广州地铁集团有限公司 Engineering quality acceptance form data structured storage method and system
CN117076454B (en) * 2023-08-21 2024-03-12 广州地铁集团有限公司 Engineering quality acceptance form data structured storage method and system

Similar Documents

Publication Publication Date Title
CN110059206A (en) A kind of extensive hashing image search method based on depth representative learning
Wang et al. Enhancing sketch-based image retrieval by cnn semantic re-ranking
CN106407352B (en) Traffic image search method based on deep learning
Cai et al. Medical image retrieval based on convolutional neural network and supervised hashing
CN109299342B (en) Cross-modal retrieval method based on cycle generation type countermeasure network
Ma et al. Multi-level correlation adversarial hashing for cross-modal retrieval
CN106227851B (en) The image search method of depth of seam division search based on depth convolutional neural networks
CN108920720B (en) Large-scale image retrieval method based on depth hash and GPU acceleration
CN105912611B (en) A kind of fast image retrieval method based on CNN
CN108280187B (en) Hierarchical image retrieval method based on depth features of convolutional neural network
CN111198959A (en) Two-stage image retrieval method based on convolutional neural network
Xia et al. Exploiting deep features for remote sensing image retrieval: A systematic investigation
Huang et al. Unconstrained multimodal multi-label learning
CN109241317A (en) Based on the pedestrian's Hash search method for measuring loss in deep learning network
CN108427740B (en) Image emotion classification and retrieval algorithm based on depth metric learning
CN110457514A (en) A kind of multi-tag image search method based on depth Hash
CN103617609B (en) Based on k-means non-linearity manifold cluster and the representative point choosing method of graph theory
Yang et al. Local label descriptor for example based semantic image labeling
CN109933682A (en) A kind of image Hash search method and system based on semanteme in conjunction with content information
CN114882351B (en) Multi-target detection and tracking method based on improved YOLO-V5s
CN112035689A (en) Zero sample image hash retrieval method based on vision-to-semantic network
Xu et al. Weakly supervised facial expression recognition via transferred DAL-CNN and active incremental learning
CN113076490B (en) Case-related microblog object-level emotion classification method based on mixed node graph
Zhong et al. Snake: Shape-aware neural 3d keypoint field
CN112149556B (en) Face attribute identification method based on deep mutual learning and knowledge transfer

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20190726

RJ01 Rejection of invention patent application after publication