US20190340744A1 - Image processing method, terminal and storge medium - Google Patents

Image processing method, terminal and storge medium Download PDF

Info

Publication number
US20190340744A1
US20190340744A1 US16/461,789 US201716461789A US2019340744A1 US 20190340744 A1 US20190340744 A1 US 20190340744A1 US 201716461789 A US201716461789 A US 201716461789A US 2019340744 A1 US2019340744 A1 US 2019340744A1
Authority
US
United States
Prior art keywords
images
target
target images
determining
threshold value
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US16/461,789
Other versions
US10467743B1 (en
Inventor
Donglu Wang
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Intellifusion Technologies Co Ltd
Original Assignee
Shenzhen Intellifusion Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Intellifusion Technologies Co Ltd filed Critical Shenzhen Intellifusion Technologies Co Ltd
Assigned to SHENZHEN INTELLIFUSION TECHNOLOGIES CO., LTD. reassignment SHENZHEN INTELLIFUSION TECHNOLOGIES CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: WANG, Donglu
Application granted granted Critical
Publication of US10467743B1 publication Critical patent/US10467743B1/en
Publication of US20190340744A1 publication Critical patent/US20190340744A1/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/16Matrix or vector computation, e.g. matrix-matrix or matrix-vector multiplication, matrix factorization
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration using two or more images, e.g. averaging or subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • G06V10/267Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/42Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • G06T2207/30201Face

Definitions

  • the present disclosure relates to image processing technical field, and more particularly, to an image processing method, a terminal and a storage medium.
  • face recognition technology has been widely used in video monitoring field.
  • the accuracy of face detection as the first step during the face recognition application has a great impact on the performance of face recognition.
  • the face detection needs to be strong robustness, because in practical applications, face images are often affected by a variety of factors, such as illumination, occlusion of light, attitude change, etc.
  • the face detection is the most frequently invoked during face recognition process, so the face detection needs to be able to be executed efficiently.
  • Face detection technology can be realized mainly based on the characteristics of manual design, such as Haar characteristics, LBP (Local Binary Patterns) histogram characteristics, HOG (histogram of oriented gradient) characteristics, etc.
  • the computing time of the characteristics can be accepted, and satisfactory results can be obtained in practical applications, so the characteristics above-mentioned are widely used.
  • it is unable to determine whether the faces in a group of images are from the same person.
  • Exemplary embodiments of the present disclosure provide an image processing method, a terminal and a storage medium, which can be determined whether the faces in a group of images are from the same person.
  • the method includes:
  • each of the target images includes a face image, and N is an integer greater than one;
  • the N target images include face images of different people when the upper information entropy limit is greater than the first pre-set threshold value
  • the N target images include face images of the same person when the upper information entropy limit is less than or equal to the first pre-set threshold value.
  • the step of determining a covariance matrix of the N target images includes:
  • the step of determining a covariance matrix of the N target images includes:
  • the step of acquiring N target images includes:
  • the step of acquiring N target images includes:
  • the N target images correspond to a target tag
  • the method further includes:
  • the method further includes:
  • the terminal includes:
  • an acquiring unit configured to acquire N target images, wherein each of the target images comprises a face image, and N is an integer greater than one;
  • a first determining unit configured to determine a covariance matrix of the N target images
  • a second determining unit configured to determine an upper information entropy limit according to the covariance matrix
  • a judging unit configured to determine whether the upper information entropy limit is greater than a first pre-set threshold value
  • a third determining unit configured to determine that the N target images include face images of different people when the judging unit 304 determines that the upper information entropy limit is greater than the first pre-set threshold value
  • the third determining unit further configured to determine that the N target images include face images of the same person when the judging unit 304 determines that the upper information entropy limit is less than or equal to the first pre-set threshold value.
  • the first determining unit includes:
  • an extracting module configured to extract feature data of each target image in the N target images respectively to obtain N feature data
  • a calculating module configured to calculate the covariance matrix according to the N feature data.
  • the first determining unit includes:
  • a first segmenting module configured to segment the N target images respectively to obtain the N face images
  • a first determining module configured to determine the covariance matrix of the N face image.
  • the acquiring unit includes:
  • a first acquiring module configured to acquire M first images, wherein M is a positive integer
  • a second segmenting module configured to perform image segmentation on the M first images to obtain the N target images.
  • the acquiring unit includes:
  • a second acquiring module configured to acquire N second images
  • a second determining module configured to determine an image quality evaluation value of each of the N second images to obtain N image quality evaluation values
  • a processing module configured to perform image enhancement processing on the second images whose image quality evaluation values are less than a second pre-set threshold value, and keeping the second images whose image quality evaluation values are greater than or equal to the second pre-set threshold value to obtain the N target images.
  • the N target images correspond to a target tag
  • the terminal further includes:
  • an assessment unit configured to evaluate the accuracy of the target tag according to the upper information entropy limit after the judging unit determines the upper information entropy limit according to the covariance matrix.
  • the terminal further includes:
  • a processing unit configured to delete the N target images when the accuracy is lower than a third pre-set threshold value, and retain the N target images when the accuracy is higher than or equal to the third pre-set threshold value after the assessment unit evaluates the accuracy of the target tag according to the upper information entropy limit.
  • the image processing method of the embodiment includes: acquiring N target images, wherein each of the target images includes a face image, and N is an integer greater than one; determining a covariance matrix of the N target images; determining an upper information entropy limit according to the covariance matrix; determining whether the upper information entropy limit is greater than a first pre-set threshold value; when the upper information entropy limit is greater than the first pre-set threshold value, determining that the N target images include face images of different people; when the upper information entropy limit is less than or equal to the first pre-set threshold value, determining that the N target images include face images of the same person.
  • FIG. 1 is a flowchart of an image processing method provided in one embodiment of the present invention
  • FIG. 2 is a flowchart of an image processing method provided in another embodiment of the present invention.
  • FIG. 3 is a block diagram of a terminal provided in one embodiment of the present invention.
  • FIG. 4 is a block diagram of a first determining unit as shown in FIG. 3 provided in one embodiment of the present invention.
  • FIG. 5 is a block diagram of a first determining unit as shown in FIG. 3 provided in another embodiment of the present invention.
  • FIG. 6 is a block diagram of an acquiring unit as shown in FIG. 3 provided in one embodiment of the present invention.
  • FIG. 7 is a block diagram of an acquiring unit as shown in FIG. 3 provided in another embodiment of the present invention.
  • FIG. 8 is a block diagram of a terminal provided in another embodiment of the present invention.
  • a terminal described in the embodiments of the disclosure may include a smart phone (such as an Android phone, a iOS phone, a Windows Phone, etc.), a tablet computer, a handheld computer, a laptop, a mobile internet device (MID) or a wearable device, etc., which is only an example, not exhaustive, and is not limited the terminals listed above.
  • a smart phone such as an Android phone, a iOS phone, a Windows Phone, etc.
  • a tablet computer such as an Android phone, a iOS phone, a Windows Phone, etc.
  • a handheld computer such as a tablet computer, a handheld computer, a laptop, a mobile internet device (MID) or a wearable device, etc.
  • MID mobile internet device
  • a metric is often used to reflect the degree of difference between data.
  • the degree of difference between the facial feature data can be used to determine whether the group of given facial feature data belong to the same person or not.
  • variance can be used as a metric to measure the degree of difference between data.
  • covariance can be calculated correspondingly. The covariance exists in the form of matrix, and cannot be directly used as a metric (scalar).
  • a metric method that can be used to measure the degree of difference between the multidimensional data is needed.
  • a method for measuring degree of difference between multidimensional data is provided in the embodiment of the disclosure, which can better retain information related to the degree of difference in a covariance matrix. Moreover, this method can reasonably explain the data of mixed Gaussian distribution in term of probability and informatics, and can be used to determine whether a set of (mixed Gaussian distribution) facial feature data belongs to the same person or not.
  • the variance is usually used, and the generalization of the variance on the multidimensional data is the covariance matrix.
  • a matrix cannot be measured as a scalar, so that a method for effectively measuring degree of difference between multidimensional data is needed.
  • the covariance matrix can be defined as follows:
  • the degree of variance between data can be estimated by the following method. Firstly, the covariance matrix ⁇ of the multidimensional data is calculated; the value det ( ⁇ ) (or any a monotone mapping, such as log (det ( ⁇ ))) of matrix determinant is used as the metric of the degree of variance between the data, and the larger the value det ( ⁇ ), the greater the degree of difference between the data.
  • det ( ⁇ ) or any a monotone mapping, such as log (det ( ⁇ )) of matrix determinant
  • the image processing method in this embodiment of the present invention includes the following steps:
  • each of the target images includes a face image, and N is an integer greater than one.
  • Each target image of the N target images can include one face image, and N is an integer greater than one.
  • the N target images can have the same or partially the same format size.
  • the format of the N target images can be the same or partially the same.
  • the step of acquiring N target images can include the following steps:
  • the image segmentation can adopt at least one of grayscale threshold segmentation, region growing, clustering analysis, wavelet transform, and maximum between-cluster variance algorithm (namely, OTSU method).
  • the step of acquiring N target images can include the following steps:
  • Each of the N second images includes a face image
  • the image quality of each second image in the N second images can be evaluated to determine the image quality evaluation value of each of the N second images, thereby obtaining the N image quality evaluation values.
  • the image quality evaluation value i in the N image quality evaluation values is less than the second pre-set threshold value
  • the second images corresponding to the image quality evaluation value i are performed the image enhancement processing, wherein the image quality evaluation value i can be any one of the N image quality evaluation values.
  • the image quality evaluation value j in the N image quality evaluation values is greater than or equal to the second pre-set threshold value, then the second images corresponding to the image quality evaluation value j are kept.
  • At least one image quality evaluation index can be used to evaluate the image quality of the second images to obtain the image quality evaluation value.
  • the image quality evaluation index can include but not limited to: average grey scale, standard deviation, entropy, edge-preserving index, SNR (signal to noise ratio), and so on.
  • image quality evaluation indexes can be used to evaluate the image quality.
  • the image quality evaluation indexes are not the more the better, that is, it is not that the more image quality evaluation indexes are used, the better the image quality will be. That is because the more the image quality evaluation are used, the higher the computational complexity during the image quality evaluation, thereby the effect of the image quality evaluation is not necessarily better.
  • 2 to 10 image quality evaluation indexes can be used to evaluate the image quality. It can be understood that the number of image quality evaluation indexes and which image quality evaluation indexes are selected should be determined according to the specific implementation situation. Of course, it is necessary to select the image quality evaluation indexes in combination with specific scenes. For example, the image quality evaluation indexes selected for image quality evaluation in dark environment can be different from those selected for image quality evaluation in bright environment.
  • an image quality evaluation index can be used to evaluate the image quality.
  • the entropy can be used to evaluate the image quality of the second image, the higher the entropy is, the better the image quality will be; on the contrary, the lower the entropy is, the worse the image quality will be.
  • multiple image quality evaluation indexes can be used to evaluate the second images.
  • the weight of each of multiple image quality evaluation indexes can be pre-set to obtain multiple image quality evaluation values, and a final image quality evaluation value can be obtained according to multiple image quality evaluation values and the weight of each image quality evaluation index.
  • three image quality evaluation indexes are: A, B and C, the weight of A is a1, the weight of B is a2, and the weight of C is a3; when the image quality evaluation indexes A, B and C are used for image quality evaluation on a certain image, the image quality evaluation value corresponding to A is b, the image quality evaluation value corresponding to B is b2, and the image quality evaluation value corresponding to C is b3, so the final image quality evaluation value is a1b1+a2b2+a3b3.
  • the step of evaluating the image quality of the second images to obtain the image quality evaluation value may include:
  • the second images may not only include target areas, any may also include other areas, therefore, part of the second images can be selected as the target areas.
  • the center of the second image can be taken as a dot, and the area within a certain range near the dot can be taken as the target area.
  • the targets exist in the target area, that is, when image recognition is carried out in the target area, the probability of getting the targets that the user wants is relatively large, therefore improving the recognition speed.
  • step B feature point extraction is performed in the target area in order to obtain the number of the feature points of the target area, that is, the feature points in the target area can be extracted to obtain the number of the feature points in the target area.
  • feature point extraction can be: Harris corner extraction, scale invariant feature transform (SIFT), etc.
  • SIFT scale invariant feature transform
  • the number of pre-set feature points can be predetermined by the user or be the system default, and the ratio between the number of feature points in the target area and the number of pre-set feature points can be taken as the image quality evaluation value.
  • the covariance matrix of the N target images can be calculated.
  • the N target images are performed for feature extraction to obtain N feature data sets, and each target image in the N target images corresponds to a feature data set.
  • the covariance matrix of the N target images can be determined according to the N feature data sets.
  • the way of the feature extraction can be, but not limited to: neural network algorithm, genetic algorithm, machine learning algorithm, Harris corner detection algorithm, etc.
  • the step of determining a covariance matrix of the N target images includes:
  • the feature data of each target image in the N target images can be extracted respectively to obtain the N Gaussian distribution feature data, that is, each target image can correspond to one of the Gaussian distribution feature data.
  • the covariance matrix can be calculated according to the N Gaussian distribution feature data.
  • the way of extracting feature data in step 21 can be: neural network algorithm, genetic algorithm, machine learning algorithm, Harris corner detection algorithm, etc.
  • the step of determining a covariance matrix of the N target images includes:
  • each target image of the N target images can not only include face image areas, can also include other areas, therefore, each target image of the N target images can be segmented to obtain the N face images.
  • one face image can be obtained in each target image, thereby obtaining N face images.
  • the feature data of each face image in the N face images can be extracted to obtain the N feature data, that is, each face image can correspond to one of the feature data.
  • the covariance matrix can be calculated according to the feature data.
  • Entropy is used to represent the degree of data chaos in informatics. Entropy is defined as:
  • Gaussian distribution is a distribution which can make the upper information entropy limit maximization. That is, when P (x) is the Gaussian distribution, the information entropy G (x) is the upper limit of the entropy H (x):
  • G (x) is a monotone function of the determinant det ( ⁇ ), that is, the larger the value of covariance matrix determinant is, the higher the upper information entropy limit will be.
  • the upper information entropy limit can be a specific value, and can be other variation forms, or an upper information entropy limit function or its characteristics (such as monotonicity), such as log(A), where A is the specific value of the upper information entropy limit.
  • step 104 can be executed under this premise.
  • Face image feature extraction method can be used to represent a face image with a multi-dimensional vector.
  • the upper information entropy limit can be used to evaluate whether these features belong to the same person or not.
  • the entropy H(X) can be directly figured out hopefully, and be determined by using the following method. When the entropy H(X) is greater than a threshold value t, it can be considered that the degree of data chaotic is high, which indicates that the feature data does not belong to the same person.
  • the feature data complies with multivariate Gaussian distribution. So for the feature data of any group of images, they comply with Gaussian distribution (including one person), or comply with Gaussian mixture distribution (including more than one person).
  • Gaussian distribution including one person
  • Gaussian mixture distribution including more than one person
  • the following alternative method can be used to calculate the upper information entropy limit G(X) of the given set of face feature data X.
  • G(X) is greater than the threshold value t
  • H(X) is less than or equal to the threshold value t.
  • H(X) ⁇ G(X) only when they belong to different people, H(X) ⁇ G(X), so it can be considered than in this case the data does not belong to the same person.
  • the first pre-set threshold value can be predetermined by the user or be the system default. Of course, in one embodiment, the first pre-set threshold value can be an empirical value.
  • the upper information entropy limit when the upper information entropy limit is greater than the first pre-set threshold value, it can be determined that the N target images include face images of different people. When the upper information entropy limit is less than or equal to the first pre-set threshold value, it can be determined that the N target images include face images of the same person.
  • each target image in the N target images is the face image
  • the N target images form a group of images.
  • the feature data of each target image in the N target images can be extracted to obtain N feature data
  • the covariance matrix of the N target images can be determined according to the N feature data.
  • the upper information entropy limit can be calculated according to the covariance matrix, and then the upper information entropy is determined to be greater than a first pre-set threshold value.
  • the upper information entropy limit is greater than the first pre-set threshold value, it can be determined that the face images in the N target images are different people.
  • the N target images can correspond to one tag, such as passer-by A.
  • passer-by A When the face images in the N target images are different people, the tag is wrong; when the face images in the N target images are the same person, the tag is correct.
  • the trace of the covariance matrix can be directly calculated, which is equivalent to the sum of the variances of each dimension.
  • the method of superposing each of the dimensions linearly would lose the correlation information among the dimensions, and there is no intuitive theoretical explanation.
  • the trace of the covariance matrix is described as follows.
  • the variance of a first dimension is a
  • the variance of a second dimension is b. If the first dimension and the second dimension are independent, then the covariance matrix is
  • the trace cannot well reflect the correlation information among the dimensions.
  • the embodiment of the invention is adopted to determine whether the face images in a group of images (such as, the above-mentioned N target images) belong to the same person, which has a higher recognition accuracy, can better retain the information related to the degree of difference in the covariance matrix, and has a reasonable explanation for mixed Gaussian distribution data in probability and informatics.
  • the method mentioned above combined with the method of face feature extraction can be used to determine whether a group of face images (mixed Gaussian distribution) belong to the same person and clear face image data set and other problems.
  • the following conditions may also be included after step 104 .
  • the N target images correspond to a target tag.
  • the image processing method further includes:
  • the upper information entropy limit can be compared with a second pre-set threshold value.
  • the upper information entropy limit is less than the second pre-set threshold value, it means that the target tag is correct; when the upper information entropy limit is greater than or equal to the second pre-set threshold value, it means that the target tag is incorrect.
  • the second pre-set threshold value is different from the first pre-set threshold value.
  • the second pre-set threshold can be the default experience value of the system, or can be set by the user.
  • the image processing method of the embodiment includes: acquiring N target images, wherein each of the target images includes a face image, and N is an integer greater than one; determining a covariance matrix of the N target images; determining an upper information entropy limit according to the covariance matrix; determining whether the upper information entropy limit is greater than a first pre-set threshold value; when the upper information entropy limit is greater than the first pre-set threshold value, determining that the N target images include face images of different people; when the upper information entropy limit is less than or equal to the first pre-set threshold value, determining that the N target images include face images of the same person.
  • it can be determined whether the faces in a group of images are from the same person.
  • FIG. 2 is a flowchart of an image processing method provided in another embodiment of the present invention.
  • the image processing method in this embodiment includes the following steps:
  • each of the target images includes a face image, an N is an integer greater than one.
  • steps 201 - 203 are the same as the steps 101 - 103 of the image processing method as described in FIG. 1 , so that the specific content of steps 201 - 203 can refer to the corresponding steps 101 - 103 in the embodiment of FIG. 1 as described above.
  • the upper information entropy limit can be compared with a second pre-set threshold value.
  • the second pre-set threshold can be the default experience value of the system, or can be set by the user.
  • the trace of the covariance matrix can be directly calculated, which is equivalent to the sum of the variances of each dimension.
  • the method of superposing each of the dimensions linearly would lose the correlation information among the dimensions, and there is no intuitive theoretical explanation.
  • the embodiment of the invention is adopted to evaluate the accuracy of the target tag in a group of images (such as, the above-mentioned N target images), which has a higher evaluation accuracy, can better retain the information related to the degree of difference in the covariance matrix, and has a reasonable explanation for mixed Gaussian distribution data in probability and informatics.
  • the method mentioned above combined with the method of face feature extraction can be used to evaluate the accuracy of the target tag in a group of images (which conform to the mixed Gaussian distribution).
  • the image processing method further includes the following steps:
  • the third pre-set threshold value can be predetermined by the user or be the system default.
  • the N target images can be equivalent to a data set, and an accuracy can be obtained after step 204 .
  • the accuracy is lower than the third pre-set threshold value, the N target images can be deleted, and when the accuracy is higher than or equal to the third pre-set threshold value, the N target images can be retained.
  • each data set can correspond to one accuracy, then there are M accuracy.
  • the data set whose accuracy is lower than the third pre-set threshold value will be deleted; if a certain accuracy rate is higher than or equal to the third pre-set threshold value, the data set corresponding to that accuracy rate will be retained.
  • the image processing method includes: acquiring N target images, where each of the target images include a face image, and N is an integer greater than one; determining a covariance matrix of the N target images; determining an upper information entropy limit according to the covariance matrix; and evaluating the accuracy of tags corresponding to a set of target images according to the upper information entropy limit. Therefore, the method can improve the accuracy of target image evaluation.
  • the following embodiments provide a device used to implement the image processing method as described below.
  • FIG. 3 is a block diagram of a terminal 100 provided in one embodiment of the present invention.
  • the terminal 100 in this embodiment includes: an acquiring unit 301 , a first determining unit 302 , a second determining unit 303 , a judging unit 304 and a third determining unit 305 .
  • an acquiring unit 301 receives a first determining unit 302 from a terminal.
  • a second determining unit 303 receives signals from a signal.
  • a judging unit 304 includes a third determining unit 305 .
  • the acquiring unit 301 is configured to acquire N target images, wherein each of the target images includes a face image, and N is an integer greater than one;
  • the first determining unit 302 is configured to determine a covariance matrix of the N target images
  • the second determining unit 303 is configured to determine an upper information entropy limit according to the covariance matrix
  • the judging unit 304 is configured to determine whether the upper information entropy limit is greater than a first pre-set threshold value
  • the third determining unit 305 is configured to determine that the N target images include face images of different people when the judging unit 304 determines that the upper information entropy limit is greater than the first pre-set threshold value;
  • the third determining unit 305 is further configured to determine that the N target images include face images of the same person when the judging unit 304 determines that the upper information entropy limit is less than or equal to the first pre-set threshold value.
  • FIG. 4 is a block diagram of a first determining unit 302 as shown in FIG. 3 provided in one embodiment.
  • the first determining unit 302 includes: an extracting module 3021 and a calculating module 3022 .
  • an extracting module 3021 includes: an extracting module 3021 and a calculating module 3022 .
  • the extracting module 3021 is configured to extract feature data of each target image in the N target images respectively to obtain N feature data;
  • the calculating module 3022 is configured to calculate the covariance matrix according to the N feature data.
  • FIG. 5 is a block diagram of a first determining unit 302 as shown in FIG. 3 .
  • the first determining unit 302 includes: a first segmenting module 3023 and a first determining module 3024 .
  • a first segmenting module 3023 includes: a first segmenting module 3023 and a first determining module 3024 .
  • the first segmenting module 3023 is configured to carry out target segmentation for the N target images respectively to obtain the N face images; that is, the first segmenting module 3023 is configured to segment the N target images respectively to obtain the N face images;
  • the first determining module 3024 is configured to determine the covariance matrix of the N face image.
  • FIG. 6 is a block diagram of an acquiring unit 301 as shown in FIG. 3 provided in one embodiment of the present invention.
  • the acquiring unit 301 includes: a first acquiring module 3011 and a second segmenting module 3012 .
  • a first acquiring module 3011 and a second segmenting module 3012 .
  • the first acquiring module 3011 is configured to acquire M first images, wherein M is a positive integer
  • the second segmenting module 3012 is configured to perform image segmentation on the M first images to obtain the N target images; that is, the second segmenting module 3012 is configured to segmenting the M first images to obtain the N target images.
  • FIG. 7 is a block diagram of an acquiring unit 301 as shown in FIG. 3 provided in another embodiment of the present invention.
  • the acquiring unit 301 includes: a second acquiring module 3013 , a second determining module 3014 and a processing module 3015 .
  • a second acquiring module 3013 includes: a second acquiring module 3013 , a second determining module 3014 and a processing module 3015 .
  • the second acquiring module 3013 is configured to acquire N second images
  • the second determining module 3014 is configured to determine an image quality evaluation value of each of the N second images to obtain N image quality evaluation values;
  • the processing module 3015 is configured to perform image enhancement processing on the second images whose image quality evaluation values are less than a second pre-set threshold value, and keeping the second images whose image quality evaluation values are greater than or equal to the second pre-set threshold value to obtain the N target images.
  • the N target images correspond to a target tag.
  • the terminal 100 further includes: an assessment unit (no shown). In detail:
  • the assessment unit is configured to evaluate the accuracy of the target tag according to the upper information entropy limit after determining an upper information entropy limit according to the covariance matrix.
  • the terminal 100 further includes: a processing unit (no shown).
  • a processing unit no shown.
  • the processing unit is configured to delete the N target images when the accuracy is lower than a third pre-set threshold value, and retain the N target images when the accuracy is higher than or equal to the third pre-set threshold value after the assessment unit evaluates the accuracy of the target tag according to the upper information entropy limit.
  • the terminal 100 is configured to acquire N target images, wherein each of the target images includes a face image, and N is an integer greater than one; determine a covariance matrix of the N target images; determine an upper information entropy limit according to the covariance matrix; determine whether the upper information entropy limit is greater than a first pre-set threshold value; when the upper information entropy limit is greater than the first pre-set threshold value, determine that the N target images include face images of different people; when the upper information entropy limit is less than or equal to the first pre-set threshold value, determine that the N target images include face images of the same person.
  • the terminal 100 it can be determined whether the faces in a group of images are from the same person.
  • FIG. 8 is a block diagram of a terminal 200 provided in another embodiment of the present invention.
  • the terminal 200 includes: at least one input device 1000 , at least one output device 2000 , at least one processor 3000 , such as a CPU, and a memory 4000 .
  • the input device 1000 , the output device 2000 , the processor 3000 and the memory 4000 are connected by a bus 5000 .
  • the input device 1000 can be a touch panel, a physical button or a mouse.
  • the output device 2000 can be a display screen.
  • the memory 4000 can be a high-speed RAM memory, can also be a non-volatile memory, such as a magnetic disk storage.
  • the memory 4000 is used to store a group of program codes, and the input device 1000 , the output device 2000 and the processor 3000 are used to call the program codes stored in the memory 4000 and perform the following operations:
  • the processor 3000 is configured to execute the program codes for:
  • each of the target images includes a face image, and N is an integer greater than one;
  • the N target images include face images of different people when the upper information entropy limit is greater than the first pre-set threshold value
  • the N target images include face images of the same person when the upper information entropy limit is less than or equal to the first pre-set threshold value.
  • the program codes for determining a covariance matrix of the N target images are executed by the processor 3000 for:
  • the program codes for determining a covariance matrix of the N target images are executed by the processor 3000 for:
  • the program codes for acquiring N target images are executed by the processor 3000 for:
  • the program codes for acquiring N target images are executed by the processor 3000 for:
  • the N target images correspond to a target tag.
  • the program codes are further executed by the processor 3000 for:
  • the program codes are further executed by the processor 3000 for:
  • the embodiment of the invention also provides a computer storage medium in which a program may be stored, and the execution of the program includes part or all steps of any of the image processing methods recorded in the embodiments.
  • embodiments of the invention may be provided as methods, devices (equipments), or computer program products. Therefore, the invention may take the form of full hardware embodiments, full software embodiments, or embodiments combining software and hardware aspects. Furthermore, the present invention may take the form of a computer program product implemented on one or more computer available storage media (including but not limited to magnetic disk storage, CD-ROM, optical memory, etc.) including computer available program codes. Computer programs are stored/distributed in suitable storage medium, provided with or as part of other hardware, and may also be distributed in other forms, such as the Internet or other wired or wireless telecommunications systems.
  • the present invention is described by reference to the method, device (equipment) and flow chart and/or block diagram of a computer program product of an embodiment of the present invention.
  • Each flow and/or box in the flowchart and/or block diagram, and the combination of flow and/or box in the flowchart and/or block diagram, shall be understood to be implemented by computer program instructions.
  • the computer program instructions can be provided to a general purpose computer, a special purpose computer, an embedded processor or processors of other programmable data processing devices to produce a machine, making the instructions executed by the processors of the computer or other programmable data processing devices to produce a device used to implement the designated function of a flow or multiple flow in a flow chart and/or a box or multiple box in a block diagram.
  • These computer program instructions can also be stored in the computer readable storage which can boot a computer or other programmable data processing equipment in a particular way to work, making the instructions stored in the computer readable storage produce manufactures including command devices, the command devices can implement the designated function of a flow or multiple flow in a flow chart and/or a box or multiple box in a block diagram.
  • the computer program instructions can be loaded into a computer or other programmable data processing equipment, making the computer or other programmable devices to perform a series of operating steps to produce the processing of computer implementation, therefore the instructions executed on a computer or other programmable devices can provide the steps for implementing the designated function of a flow or multiple flow in a flow chart and/or a box or multiple box in a block diagram.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Computational Mathematics (AREA)
  • Pure & Applied Mathematics (AREA)
  • Computing Systems (AREA)
  • Algebra (AREA)
  • Quality & Reliability (AREA)
  • Databases & Information Systems (AREA)
  • Software Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

An image processing method, a terminal and a storage medium. The method includes: acquiring N target image, wherein each of the target images contains a face image, and N is an integer greater than one (101); determining a covariance matrix of the N target image (102); determining an upper information entropy limit according to the covariance matrix (103); determining whether the upper information entropy limit is greater than a first pre-set threshold value (104); when the upper information entropy limit is greater than the first pre-set threshold value, determining that the N target images contain face images of different people (105); and when the upper information entropy limit is less than or equal to the first pre-set threshold value, determining that the N target images contain face images of the same person (106).

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application claims the benefit of priority from Chinese Patent Application NO. 201611026182.5 entitled “AN IMAGE PROCESSING METHOD AND A TERMINAL” and filed on Nov. 18, 2016, the content of which is hereby incorporated in its entire by reference.
  • FIELD
  • The present disclosure relates to image processing technical field, and more particularly, to an image processing method, a terminal and a storage medium.
  • BACKGROUND
  • With the rapid development of information technology, face recognition technology has been widely used in video monitoring field. In the field of face recognition application, the accuracy of face detection, as the first step during the face recognition application has a great impact on the performance of face recognition. The face detection needs to be strong robustness, because in practical applications, face images are often affected by a variety of factors, such as illumination, occlusion of light, attitude change, etc. The face detection is the most frequently invoked during face recognition process, so the face detection needs to be able to be executed efficiently. Face detection technology can be realized mainly based on the characteristics of manual design, such as Haar characteristics, LBP (Local Binary Patterns) histogram characteristics, HOG (histogram of oriented gradient) characteristics, etc. The computing time of the characteristics can be accepted, and satisfactory results can be obtained in practical applications, so the characteristics above-mentioned are widely used. However, in the existing technology, it is unable to determine whether the faces in a group of images are from the same person.
  • SUMMARY
  • Exemplary embodiments of the present disclosure provide an image processing method, a terminal and a storage medium, which can be determined whether the faces in a group of images are from the same person.
  • According to a first aspect of the invention, it provides an image processing method, the method includes:
  • acquiring N target images, wherein each of the target images includes a face image, and N is an integer greater than one;
  • determining a covariance matrix of the N target images;
  • determining an upper information entropy limit according to the covariance matrix;
  • determining whether the upper information entropy limit is greater than a first pre-set threshold value;
  • determining that the N target images include face images of different people when the upper information entropy limit is greater than the first pre-set threshold value; and
  • determining that the N target images include face images of the same person when the upper information entropy limit is less than or equal to the first pre-set threshold value.
  • In combination with the first aspect, in a first possible embodiment of the first aspect, the step of determining a covariance matrix of the N target images, includes:
  • extracting feature data of each target image in the N target images respectively to obtain N feature data;
  • calculating the covariance matrix according to the N feature data.
  • In combination with the first aspect, in a second possible embodiment of the first aspect, the step of determining a covariance matrix of the N target images, includes:
  • carrying out target segmentation for the N target images respectively to obtain the N face images;
  • determining the covariance matrix of the N face images.
  • In combination with the first aspect or the first or the second possible embodiment of the first aspect, in a third possible embodiment of the first aspect, the step of acquiring N target images, includes:
  • acquiring N second images;
  • determining an image quality evaluation value of each of the N second images to obtain N image quality evaluation values;
  • performing image enhancement processing on the second images whose image quality evaluation values are less than a second pre-set threshold value, and keeping the second images whose image quality evaluation values are greater than or equal to the second pre-set threshold value to obtain the N target images.
  • In combination with the first aspect or the first or the second possible embodiment of the first aspect, in a fourth possible embodiment of the first aspect, the step of acquiring N target images, includes:
  • acquiring M first images, wherein M is a positive integer;
  • performing image segmentation on the M first images to obtain the N target images.
  • In combination with the first aspect or the first or the second possible embodiment of the first aspect, in a fifth possible embodiment of the first aspect, the N target images correspond to a target tag;
  • after the step of determining an upper information entropy limit according to the covariance matrix, the method further includes:
  • evaluating the accuracy of the target tag according to the upper information entropy limit.
  • In combination with the fifth possible of the first aspect, in a sixth possible embodiment of the first aspect, after the step of evaluating the accuracy of the target tag according to the upper information entropy limit, the method further includes:
  • deleting the N target images when the accuracy is lower than a third pre-set threshold value;
  • retaining the N target images when the accuracy is higher than or equal to the third pre-set threshold value.
  • According to a second aspect of the invention, it provides a terminal, the terminal includes:
  • an acquiring unit, configured to acquire N target images, wherein each of the target images comprises a face image, and N is an integer greater than one;
  • a first determining unit, configured to determine a covariance matrix of the N target images;
  • a second determining unit, configured to determine an upper information entropy limit according to the covariance matrix;
  • a judging unit, configured to determine whether the upper information entropy limit is greater than a first pre-set threshold value;
  • a third determining unit, configured to determine that the N target images include face images of different people when the judging unit 304 determines that the upper information entropy limit is greater than the first pre-set threshold value;
  • the third determining unit, further configured to determine that the N target images include face images of the same person when the judging unit 304 determines that the upper information entropy limit is less than or equal to the first pre-set threshold value.
  • In combination with the second aspect, in a first possible embodiment of the second aspect, the first determining unit includes:
  • an extracting module, configured to extract feature data of each target image in the N target images respectively to obtain N feature data;
  • a calculating module, configured to calculate the covariance matrix according to the N feature data.
  • In combination with the second aspect, in a second possible embodiment of the second aspect, the first determining unit includes:
  • a first segmenting module, configured to segment the N target images respectively to obtain the N face images;
  • a first determining module configured to determine the covariance matrix of the N face image.
  • In combination with the second aspect or the first or the second possible embodiment of the second aspect, in a third possible embodiment of the second aspect, the acquiring unit includes:
  • a first acquiring module, configured to acquire M first images, wherein M is a positive integer;
  • a second segmenting module, configured to perform image segmentation on the M first images to obtain the N target images.
  • In combination with the second aspect or the first or the second possible embodiment of the second aspect, in a fourth possible embodiment of the second aspect, the acquiring unit includes:
  • a second acquiring module, configured to acquire N second images;
  • a second determining module, configured to determine an image quality evaluation value of each of the N second images to obtain N image quality evaluation values;
  • a processing module, configured to perform image enhancement processing on the second images whose image quality evaluation values are less than a second pre-set threshold value, and keeping the second images whose image quality evaluation values are greater than or equal to the second pre-set threshold value to obtain the N target images.
  • In combination with the second aspect or the first or the second possible embodiment of the second aspect, in a fifth possible embodiment of the second aspect, the N target images correspond to a target tag;
  • the terminal further includes:
  • an assessment unit, configured to evaluate the accuracy of the target tag according to the upper information entropy limit after the judging unit determines the upper information entropy limit according to the covariance matrix.
  • In combination with the fifth possible embodiment of the second aspect, in a sixth possible embodiment of the second aspect, the terminal further includes:
  • a processing unit, configured to delete the N target images when the accuracy is lower than a third pre-set threshold value, and retain the N target images when the accuracy is higher than or equal to the third pre-set threshold value after the assessment unit evaluates the accuracy of the target tag according to the upper information entropy limit.
  • The embodiments of the disclosure have the following beneficial effects:
  • the image processing method of the embodiment includes: acquiring N target images, wherein each of the target images includes a face image, and N is an integer greater than one; determining a covariance matrix of the N target images; determining an upper information entropy limit according to the covariance matrix; determining whether the upper information entropy limit is greater than a first pre-set threshold value; when the upper information entropy limit is greater than the first pre-set threshold value, determining that the N target images include face images of different people; when the upper information entropy limit is less than or equal to the first pre-set threshold value, determining that the N target images include face images of the same person. By means of the method, it can be determined whether the faces in a group of images are from the same person.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • In order to more clearly understand the technical solution hereinafter in embodiments of the present disclosure, a brief description to the drawings used in detailed description of embodiments hereinafter is provided thereof. Obviously, the drawings described below are some embodiments of the invention, for persons of ordinary skills in this field, other drawings can be obtained according to the drawings below on the premise of no creative work.
  • FIG. 1 is a flowchart of an image processing method provided in one embodiment of the present invention;
  • FIG. 2 is a flowchart of an image processing method provided in another embodiment of the present invention;
  • FIG. 3 is a block diagram of a terminal provided in one embodiment of the present invention;
  • FIG. 4 is a block diagram of a first determining unit as shown in FIG. 3 provided in one embodiment of the present invention;
  • FIG. 5 is a block diagram of a first determining unit as shown in FIG. 3 provided in another embodiment of the present invention;
  • FIG. 6 is a block diagram of an acquiring unit as shown in FIG. 3 provided in one embodiment of the present invention;
  • FIG. 7 is a block diagram of an acquiring unit as shown in FIG. 3 provided in another embodiment of the present invention;
  • FIG. 8 is a block diagram of a terminal provided in another embodiment of the present invention.
  • DETAILED DESCRIPTION
  • Reference will now be made in detail to embodiments, examples of which are illustrated in the accompanying drawings. In the following detailed description, numerous specific details are set forth in order to provide a thorough understanding of the subject matter presented herein. But it will be apparent to one skilled in the art that the subject matter may be practiced without these specific details. Based on the embodiments of the invention, all other embodiments obtained by persons of ordinary skills in this field without creative work shall fall within the protection scope of the present disclosure.
  • The terms “first”, “second”, “third” and “fourth” in specification, claims and drawings of the present disclosure are used to distinguish different objects, but not used to describe a particular sequence. In addition, the terms “include” and “have” and any deformation thereof are intended to cover exclusive inclusions. For example, a process, method, system, product or device that contains a series of steps or units is not only limited to listed steps or units, but optionally includes steps or units not listed, or optionally includes other inherent steps or units for the processes, methods, products or devices.
  • The reference to “embodiments” in this invention means specific characteristics, structures or characters described in the embodiments can be included in at least one embodiment of the present invention. The term “embodiment” shown in various positions in the specification does not necessarily refer to the same embodiment, also does not refer to the embodiments that are independent and exclusive embodiments with other embodiments or alternative embodiments. It can be understood both explicitly and implicitly by persons of ordinary skills in this field that the embodiments described herein can be combined with other embodiment.
  • A terminal described in the embodiments of the disclosure may include a smart phone (such as an Android phone, a iOS phone, a Windows Phone, etc.), a tablet computer, a handheld computer, a laptop, a mobile internet device (MID) or a wearable device, etc., which is only an example, not exhaustive, and is not limited the terminals listed above.
  • It should be noted that, in practical applications, a metric (scalar) is often used to reflect the degree of difference between data. For example, for a group of given facial feature data, the degree of difference between the facial feature data can be used to determine whether the group of given facial feature data belong to the same person or not. For one-dimensional data, variance can be used as a metric to measure the degree of difference between data. For multidimensional data, covariance can be calculated correspondingly. The covariance exists in the form of matrix, and cannot be directly used as a metric (scalar). Thus, a metric method that can be used to measure the degree of difference between the multidimensional data is needed.
  • A method for measuring degree of difference between multidimensional data is provided in the embodiment of the disclosure, which can better retain information related to the degree of difference in a covariance matrix. Moreover, this method can reasonably explain the data of mixed Gaussian distribution in term of probability and informatics, and can be used to determine whether a set of (mixed Gaussian distribution) facial feature data belongs to the same person or not.
  • For the one-dimensional data, the variance is usually used, and the generalization of the variance on the multidimensional data is the covariance matrix. A matrix cannot be measured as a scalar, so that a method for effectively measuring degree of difference between multidimensional data is needed.
  • For example, the definition of variance can be expressed as follows:

  • Var(X)=E[(X−μ)2].
  • For the multidimensional data, the covariance matrix can be defined as follows:
  • Σ = [ E [ ( X 1 - μ 1 ) ( X 1 - μ 1 ) ] E [ ( X 1 - μ 1 ) ( X 2 - μ 2 ) ] E [ ( X 1 - μ 1 ) ( X n - μ n ) ] E [ ( X 2 - μ 2 ) ( X 1 - μ 1 ) ] E [ ( X 2 - μ 2 ) ( X 2 - μ 2 ) ] E [ ( X 2 - μ 2 ) ( X n - μ n ) ] E [ ( X n - μ n ) ( X 1 - μ 1 ) ] E [ ( X n - μ n ) ( X 2 - μ 2 ) ] E [ ( X n - μ n ) ( X n - μ n ) ] ]
  • For a set of multidimensional data X, the degree of variance between data can be estimated by the following method. Firstly, the covariance matrix Σ of the multidimensional data is calculated; the value det (Σ) (or any a monotone mapping, such as log (det (Σ))) of matrix determinant is used as the metric of the degree of variance between the data, and the larger the value det (Σ), the greater the degree of difference between the data. The theoretical explanation of this method and how to apply it to facial feature data based on mixed Gaussian model are given below.
  • Referring to FIG. 1, a flowchart of an image processing method provided in one embodiment of the present invention is described in detail as below. The image processing method in this embodiment of the present invention includes the following steps:
  • 101, acquiring N target images, wherein each of the target images includes a face image, and N is an integer greater than one.
  • Each target image of the N target images can include one face image, and N is an integer greater than one.
  • Optionally, the N target images can have the same or partially the same format size.
  • Optionally, the format of the N target images can be the same or partially the same.
  • Optionally, the step of acquiring N target images can include the following steps:
  • 11), acquiring M first images, wherein M is a positive integer;
  • 12), performing image segmentation on the M first images to obtain the N target images.
  • In detail, M can be one or an integer greater than one, and when M is equal to one, it can be understood that one first image can include N target images. When M is greater than one, it can be understood that M first images include N target images. Understandably, the relationship between M and N can be: M greater than N, M=N, or M less than N. for example, when M is greater than N, it means that not every first image in M first images includes the face image. When M is equal to N, it means that each first image in M first images may include a face image. When M is less than N, it means that at least one first image include more than two face images. In a word, the M first images can include N face images. Therefore, the M first images can be segmented respectively, to acquire the N target images. Each target image in the N target images include a face image.
  • The image segmentation can adopt at least one of grayscale threshold segmentation, region growing, clustering analysis, wavelet transform, and maximum between-cluster variance algorithm (namely, OTSU method).
  • In one embodiment, the step of acquiring N target images can include the following steps:
  • 13), acquiring N second images;
  • 14), determining an image quality evaluation value of each of the N second images to obtain N image quality evaluation values;
  • 15), performing image enhancement processing on the second images whose image quality evaluation values are less than a second pre-set threshold value, and keeping the second images whose image quality evaluation values are greater than or equal to the second pre-set threshold value to obtain the N target images.
  • Each of the N second images includes a face image, the image quality of each second image in the N second images can be evaluated to determine the image quality evaluation value of each of the N second images, thereby obtaining the N image quality evaluation values. Furthermore, when the image quality evaluation value i in the N image quality evaluation values is less than the second pre-set threshold value, then the second images corresponding to the image quality evaluation value i are performed the image enhancement processing, wherein the image quality evaluation value i can be any one of the N image quality evaluation values. When the image quality evaluation value j in the N image quality evaluation values is greater than or equal to the second pre-set threshold value, then the second images corresponding to the image quality evaluation value j are kept.
  • Optionally, at least one image quality evaluation index can be used to evaluate the image quality of the second images to obtain the image quality evaluation value. The image quality evaluation index can include but not limited to: average grey scale, standard deviation, entropy, edge-preserving index, SNR (signal to noise ratio), and so on.
  • Understandably, due to the limitation of using a single quality evaluation index to evaluate image quality, therefore, multiple image quality evaluation indexes can be used to evaluate the image quality. Of course, when the image quality is evaluated, the image quality evaluation indexes are not the more the better, that is, it is not that the more image quality evaluation indexes are used, the better the image quality will be. That is because the more the image quality evaluation are used, the higher the computational complexity during the image quality evaluation, thereby the effect of the image quality evaluation is not necessarily better. Thus, in the case of higher requirements for the image quality evaluation, 2 to 10 image quality evaluation indexes can be used to evaluate the image quality. It can be understood that the number of image quality evaluation indexes and which image quality evaluation indexes are selected should be determined according to the specific implementation situation. Of course, it is necessary to select the image quality evaluation indexes in combination with specific scenes. For example, the image quality evaluation indexes selected for image quality evaluation in dark environment can be different from those selected for image quality evaluation in bright environment.
  • In one embodiment, in the case of low requirements for image quality evaluation accuracy, an image quality evaluation index can be used to evaluate the image quality. For example, the entropy can be used to evaluate the image quality of the second image, the higher the entropy is, the better the image quality will be; on the contrary, the lower the entropy is, the worse the image quality will be.
  • In one embodiment, in the case of high requirements for image quality evaluation accuracy, multiple image quality evaluation indexes can be used to evaluate the second images. When multiple image quality evaluation indexes are used for image quality evaluation of the second images, the weight of each of multiple image quality evaluation indexes can be pre-set to obtain multiple image quality evaluation values, and a final image quality evaluation value can be obtained according to multiple image quality evaluation values and the weight of each image quality evaluation index. For example, three image quality evaluation indexes are: A, B and C, the weight of A is a1, the weight of B is a2, and the weight of C is a3; when the image quality evaluation indexes A, B and C are used for image quality evaluation on a certain image, the image quality evaluation value corresponding to A is b, the image quality evaluation value corresponding to B is b2, and the image quality evaluation value corresponding to C is b3, so the final image quality evaluation value is a1b1+a2b2+a3b3. Generally, the higher the image quality evaluation value is, the better the image quality will be.
  • Optionally, the step of evaluating the image quality of the second images to obtain the image quality evaluation value, may include:
  • A), selecting face image areas in the second images;
  • B), determining the number of feature points in the face image areas;
  • C), comparing the number of feature points with the number of pre-set feature points to obtain a ratio, and taking the ratio as the image quality evaluation value of the second image.
  • In detail, because the second images may not only include target areas, any may also include other areas, therefore, part of the second images can be selected as the target areas. For example, the center of the second image can be taken as a dot, and the area within a certain range near the dot can be taken as the target area. There is a high probability that the targets exist in the target area, that is, when image recognition is carried out in the target area, the probability of getting the targets that the user wants is relatively large, therefore improving the recognition speed. In step B, feature point extraction is performed in the target area in order to obtain the number of the feature points of the target area, that is, the feature points in the target area can be extracted to obtain the number of the feature points in the target area. In one embodiment, feature point extraction can be: Harris corner extraction, scale invariant feature transform (SIFT), etc. The number of pre-set feature points can be predetermined by the user or be the system default, and the ratio between the number of feature points in the target area and the number of pre-set feature points can be taken as the image quality evaluation value.
  • 102, determining a covariance matrix of the N target images.
  • In detail, the covariance matrix of the N target images can be calculated. Specifically, the N target images are performed for feature extraction to obtain N feature data sets, and each target image in the N target images corresponds to a feature data set. Thus, the covariance matrix of the N target images can be determined according to the N feature data sets. The way of the feature extraction can be, but not limited to: neural network algorithm, genetic algorithm, machine learning algorithm, Harris corner detection algorithm, etc.
  • Optionally, the step of determining a covariance matrix of the N target images, includes:
  • 21), extracting feature data of each target image in the N target images respectively to obtain N Gaussian distribution feature data;
  • 22), calculating the covariance matrix according to the N Gaussian distribution feature data.
  • The feature data of each target image in the N target images can be extracted respectively to obtain the N Gaussian distribution feature data, that is, each target image can correspond to one of the Gaussian distribution feature data. The covariance matrix can be calculated according to the N Gaussian distribution feature data. The way of extracting feature data in step 21 can be: neural network algorithm, genetic algorithm, machine learning algorithm, Harris corner detection algorithm, etc.
  • Optionally, the step of determining a covariance matrix of the N target images, includes:
  • 23), carrying out target segmentation for the N target images respectively to obtain the N face images;
  • 24), determining the covariance matrix of the N face images.
  • In detail, each target image of the N target images can not only include face image areas, can also include other areas, therefore, each target image of the N target images can be segmented to obtain the N face images. Thus, one face image can be obtained in each target image, thereby obtaining N face images. The feature data of each face image in the N face images can be extracted to obtain the N feature data, that is, each face image can correspond to one of the feature data. The covariance matrix can be calculated according to the feature data.
  • 103, determining an upper information entropy limit according to the covariance matrix.
  • In general, entropy is used to represent the degree of data chaos in informatics. Entropy is defined as:

  • H(X)=∫P(x)I(x)dx=−∫P(x)logb P(x)dx,
  • wherein, P(x) represents the distribution of data, for a given covariance matrix Σ, Gaussian distribution is a distribution which can make the upper information entropy limit maximization. That is, when P (x) is the Gaussian distribution, the information entropy G (x) is the upper limit of the entropy H (x):

  • G(X)=½ln{(2πe)N det(Σ)}≥H(X),
  • wherein, G (x) is a monotone function of the determinant det (Σ), that is, the larger the value of covariance matrix determinant is, the higher the upper information entropy limit will be.
  • Optionally, the upper information entropy limit can be a specific value, and can be other variation forms, or an upper information entropy limit function or its characteristics (such as monotonicity), such as log(A), where A is the specific value of the upper information entropy limit.
  • After step 103, it can be assumed that each feature data of the N feature data satisfies the Gaussian distribution, then step 104 can be executed under this premise.
  • For face image data, a common problem is, for a given set of images, how to determine whether the set of images belongs to the same person. Different images are not suitable for being directly analyzed due to their different resolutions. Face image feature extraction method can be used to represent a face image with a multi-dimensional vector. Similarly, the upper information entropy limit can be used to evaluate whether these features belong to the same person or not. In an ideal situation, the entropy H(X) can be directly figured out hopefully, and be determined by using the following method. When the entropy H(X) is greater than a threshold value t, it can be considered that the degree of data chaotic is high, which indicates that the feature data does not belong to the same person.
  • In practice, however, it is difficult to know the true distribution of P(x), so the above method cannot be directly applied. Suppose that for different images of the same person, the feature data complies with multivariate Gaussian distribution. So for the feature data of any group of images, they comply with Gaussian distribution (including one person), or comply with Gaussian mixture distribution (including more than one person). When all the images belong to the same person, then H(X)=G(X); when the images belong to more than one person, then H(X)<G(X).
  • Therefore, for a given set of face feature data X, the following alternative method can be used to calculate the upper information entropy limit G(X) of the given set of face feature data X. When G(X) is greater than the threshold value t, it can be considered that this given set of face feature data X does not belong to the same person. The difference between this method and the method based on H(X) is: because G(X) is greater than or equal to H(X), so when G(X) is greater than the threshold value t, it is possible that H(X) is less than or equal to the threshold value t. but according to the previous assumption, only when they belong to different people, H(X)<G(X), so it can be considered than in this case the data does not belong to the same person. Thus, under the assumption of “the feature data of different images of the same person complies with multivariate Gaussian distribution”, this method is equivalent to or even superior to the method based on the real information entropy of the data.
  • 104, determining whether the upper information entropy limit is greater than a first pre-set threshold value.
  • The first pre-set threshold value can be predetermined by the user or be the system default. Of course, in one embodiment, the first pre-set threshold value can be an empirical value.
  • 105, determining that the N target images include face images of different people when the upper information entropy limit is greater than the first pre-set threshold value.
  • 106, determining that the N target images include face images of the same person when the upper information entropy limit is less than or equal to the first pre-set threshold value.
  • In detail, when the upper information entropy limit is greater than the first pre-set threshold value, it can be determined that the N target images include face images of different people. When the upper information entropy limit is less than or equal to the first pre-set threshold value, it can be determined that the N target images include face images of the same person.
  • For example, in practice, when each target image in the N target images is the face image, then the N target images form a group of images. The feature data of each target image in the N target images can be extracted to obtain N feature data, then the covariance matrix of the N target images can be determined according to the N feature data. The upper information entropy limit can be calculated according to the covariance matrix, and then the upper information entropy is determined to be greater than a first pre-set threshold value. When the upper information entropy limit is greater than the first pre-set threshold value, it can be determined that the face images in the N target images are different people. When the upper information entropy limit is less than or equal to the first pre-set threshold value, it can be determined that the face images in the N target images are the same person. Thus, the N target images can correspond to one tag, such as passer-by A. When the face images in the N target images are different people, the tag is wrong; when the face images in the N target images are the same person, the tag is correct.
  • In one embodiment, after step 102, the trace of the covariance matrix can be directly calculated, which is equivalent to the sum of the variances of each dimension. The method of superposing each of the dimensions linearly would lose the correlation information among the dimensions, and there is no intuitive theoretical explanation. The trace of the covariance matrix is described as follows.
  • For example, assuming that there is a set of two-dimensional data, the variance of a first dimension is a, and the variance of a second dimension is b. If the first dimension and the second dimension are independent, then the covariance matrix is
  • a 0 0 b ,
  • and the trace if the covariance matrix is a+b. If the first dimension and the second dimension are related to each other, then the elements besides the diagonal will no longer be equal to zero, but at the same time, the trace if the covariance matrix is still a+b. Therefore, the trace cannot well reflect the correlation information among the dimensions.
  • Moreover, there is no intuitive probability or informatics explanation for linear addition of variances of each dimension (even for independent data among dimensions). Thus, it is difficult to determine the applicability and limitations of this method.
  • In one embodiment, the embodiment of the invention is adopted to determine whether the face images in a group of images (such as, the above-mentioned N target images) belong to the same person, which has a higher recognition accuracy, can better retain the information related to the degree of difference in the covariance matrix, and has a reasonable explanation for mixed Gaussian distribution data in probability and informatics. The method mentioned above combined with the method of face feature extraction can be used to determine whether a group of face images (mixed Gaussian distribution) belong to the same person and clear face image data set and other problems.
  • In one embodiment, the following conditions may also be included after step 104. For example, the N target images correspond to a target tag. After the step 103 of determining an upper information entropy limit according to the covariance matrix, the image processing method further includes:
  • evaluating the accuracy of the target tag according to the upper information entropy limit.
  • In detail, the upper information entropy limit can be compared with a second pre-set threshold value. When the upper information entropy limit is less than the second pre-set threshold value, it means that the target tag is correct; when the upper information entropy limit is greater than or equal to the second pre-set threshold value, it means that the target tag is incorrect. It should be understood that the second pre-set threshold value is different from the first pre-set threshold value. Generally, the second pre-set threshold can be the default experience value of the system, or can be set by the user.
  • The image processing method of the embodiment includes: acquiring N target images, wherein each of the target images includes a face image, and N is an integer greater than one; determining a covariance matrix of the N target images; determining an upper information entropy limit according to the covariance matrix; determining whether the upper information entropy limit is greater than a first pre-set threshold value; when the upper information entropy limit is greater than the first pre-set threshold value, determining that the N target images include face images of different people; when the upper information entropy limit is less than or equal to the first pre-set threshold value, determining that the N target images include face images of the same person. By means of the method, it can be determined whether the faces in a group of images are from the same person.
  • Referring to FIG. 2, FIG. 2 is a flowchart of an image processing method provided in another embodiment of the present invention. The image processing method in this embodiment includes the following steps:
  • 201, acquiring N target images, wherein each of the target images includes a face image, an N is an integer greater than one.
  • 202, determining a covariance matrix of the N target images.
  • 203, determining an upper information entropy limit according to the covariance matrix.
  • The above steps 201-203 are the same as the steps 101-103 of the image processing method as described in FIG. 1, so that the specific content of steps 201-203 can refer to the corresponding steps 101-103 in the embodiment of FIG. 1 as described above.
  • 204, evaluating the accuracy of the target tag according to the upper information entropy limit.
  • In detail, the upper information entropy limit can be compared with a second pre-set threshold value. When the upper information entropy limit is less than the second pre-set threshold value, it means that the target tag is correct; when the upper information entropy limit is greater than or equal to the second pre-set threshold value, it means that the target tag is incorrect. Generally, the second pre-set threshold can be the default experience value of the system, or can be set by the user.
  • In one embodiment, the trace of the covariance matrix can be directly calculated, which is equivalent to the sum of the variances of each dimension. The method of superposing each of the dimensions linearly would lose the correlation information among the dimensions, and there is no intuitive theoretical explanation. However, the embodiment of the invention is adopted to evaluate the accuracy of the target tag in a group of images (such as, the above-mentioned N target images), which has a higher evaluation accuracy, can better retain the information related to the degree of difference in the covariance matrix, and has a reasonable explanation for mixed Gaussian distribution data in probability and informatics. The method mentioned above combined with the method of face feature extraction can be used to evaluate the accuracy of the target tag in a group of images (which conform to the mixed Gaussian distribution).
  • In one embodiment, after the step 204, the image processing method further includes the following steps:
  • deleting the N target images when the accuracy is lower than a third pre-set threshold value;
  • retaining the N target images when the accuracy is higher than or equal to the third pre-set threshold value.
  • In detail, the third pre-set threshold value can be predetermined by the user or be the system default. The N target images can be equivalent to a data set, and an accuracy can be obtained after step 204. When the accuracy is lower than the third pre-set threshold value, the N target images can be deleted, and when the accuracy is higher than or equal to the third pre-set threshold value, the N target images can be retained. In practical application, if there are M data sets, each data set can correspond to one accuracy, then there are M accuracy. In the M accuracy, if a certain accuracy is lower than the third pre-set threshold value, the data set whose accuracy is lower than the third pre-set threshold value will be deleted; if a certain accuracy rate is higher than or equal to the third pre-set threshold value, the data set corresponding to that accuracy rate will be retained.
  • In this embodiment, the image processing method includes: acquiring N target images, where each of the target images include a face image, and N is an integer greater than one; determining a covariance matrix of the N target images; determining an upper information entropy limit according to the covariance matrix; and evaluating the accuracy of tags corresponding to a set of target images according to the upper information entropy limit. Therefore, the method can improve the accuracy of target image evaluation.
  • The following embodiments provide a device used to implement the image processing method as described below.
  • Referring to FIG. 3, FIG. 3 is a block diagram of a terminal 100 provided in one embodiment of the present invention. The terminal 100 in this embodiment includes: an acquiring unit 301, a first determining unit 302, a second determining unit 303, a judging unit 304 and a third determining unit 305. In detail:
  • the acquiring unit 301 is configured to acquire N target images, wherein each of the target images includes a face image, and N is an integer greater than one;
  • the first determining unit 302 is configured to determine a covariance matrix of the N target images;
  • the second determining unit 303 is configured to determine an upper information entropy limit according to the covariance matrix;
  • the judging unit 304 is configured to determine whether the upper information entropy limit is greater than a first pre-set threshold value;
  • the third determining unit 305 is configured to determine that the N target images include face images of different people when the judging unit 304 determines that the upper information entropy limit is greater than the first pre-set threshold value;
  • the third determining unit 305 is further configured to determine that the N target images include face images of the same person when the judging unit 304 determines that the upper information entropy limit is less than or equal to the first pre-set threshold value.
  • In one embodiment, referring to FIG. 4, FIG. 4 is a block diagram of a first determining unit 302 as shown in FIG. 3 provided in one embodiment. The first determining unit 302 includes: an extracting module 3021 and a calculating module 3022. In detail:
  • the extracting module 3021 is configured to extract feature data of each target image in the N target images respectively to obtain N feature data;
  • the calculating module 3022 is configured to calculate the covariance matrix according to the N feature data.
  • In one embodiment, referring to FIG. 5, FIG. 5 is a block diagram of a first determining unit 302 as shown in FIG. 3. In this embodiment, The first determining unit 302 includes: a first segmenting module 3023 and a first determining module 3024. In detail:
  • the first segmenting module 3023 is configured to carry out target segmentation for the N target images respectively to obtain the N face images; that is, the first segmenting module 3023 is configured to segment the N target images respectively to obtain the N face images;
  • the first determining module 3024 is configured to determine the covariance matrix of the N face image.
  • Referring to FIG. 6, FIG. 6 is a block diagram of an acquiring unit 301 as shown in FIG. 3 provided in one embodiment of the present invention. In this embodiment, the acquiring unit 301 includes: a first acquiring module 3011 and a second segmenting module 3012. In detail:
  • the first acquiring module 3011 is configured to acquire M first images, wherein M is a positive integer;
  • the second segmenting module 3012 is configured to perform image segmentation on the M first images to obtain the N target images; that is, the second segmenting module 3012 is configured to segmenting the M first images to obtain the N target images.
  • Referring to FIG. 7, FIG. 7 is a block diagram of an acquiring unit 301 as shown in FIG. 3 provided in another embodiment of the present invention. In this embodiment, the acquiring unit 301 includes: a second acquiring module 3013, a second determining module 3014 and a processing module 3015. In detail:
  • the second acquiring module 3013 is configured to acquire N second images;
  • the second determining module 3014 is configured to determine an image quality evaluation value of each of the N second images to obtain N image quality evaluation values;
  • the processing module 3015 is configured to perform image enhancement processing on the second images whose image quality evaluation values are less than a second pre-set threshold value, and keeping the second images whose image quality evaluation values are greater than or equal to the second pre-set threshold value to obtain the N target images.
  • In one embodiment, the N target images correspond to a target tag. The terminal 100 further includes: an assessment unit (no shown). In detail:
  • the assessment unit is configured to evaluate the accuracy of the target tag according to the upper information entropy limit after determining an upper information entropy limit according to the covariance matrix.
  • In one embodiment, the terminal 100 further includes: a processing unit (no shown). In detail:
  • the processing unit is configured to delete the N target images when the accuracy is lower than a third pre-set threshold value, and retain the N target images when the accuracy is higher than or equal to the third pre-set threshold value after the assessment unit evaluates the accuracy of the target tag according to the upper information entropy limit.
  • The terminal 100 is configured to acquire N target images, wherein each of the target images includes a face image, and N is an integer greater than one; determine a covariance matrix of the N target images; determine an upper information entropy limit according to the covariance matrix; determine whether the upper information entropy limit is greater than a first pre-set threshold value; when the upper information entropy limit is greater than the first pre-set threshold value, determine that the N target images include face images of different people; when the upper information entropy limit is less than or equal to the first pre-set threshold value, determine that the N target images include face images of the same person. By means of the terminal 100, it can be determined whether the faces in a group of images are from the same person.
  • Referring to FIG. 8, FIG. 8 is a block diagram of a terminal 200 provided in another embodiment of the present invention. In this embodiment, the terminal 200 includes: at least one input device 1000, at least one output device 2000, at least one processor 3000, such as a CPU, and a memory 4000. The input device 1000, the output device 2000, the processor 3000 and the memory 4000 are connected by a bus 5000.
  • In this embodiment, the input device 1000 can be a touch panel, a physical button or a mouse.
  • The output device 2000 can be a display screen.
  • The memory 4000 can be a high-speed RAM memory, can also be a non-volatile memory, such as a magnetic disk storage. The memory 4000 is used to store a group of program codes, and the input device 1000, the output device 2000 and the processor 3000 are used to call the program codes stored in the memory 4000 and perform the following operations:
  • The processor 3000 is configured to execute the program codes for:
  • acquiring N target images, wherein each of the target images includes a face image, and N is an integer greater than one;
  • determining a covariance matrix of the N target images;
  • determining an upper information entropy limit according to the covariance matrix;
  • determining whether the upper information entropy limit is greater than a first pre-set threshold value;
  • determining that the N target images include face images of different people when the upper information entropy limit is greater than the first pre-set threshold value; and
  • determining that the N target images include face images of the same person when the upper information entropy limit is less than or equal to the first pre-set threshold value.
  • In one embodiment, the program codes for determining a covariance matrix of the N target images are executed by the processor 3000 for:
  • extracting feature data of each target image in the N target images respectively to obtain N feature data; and
  • calculating the covariance matrix according to the N feature data.
  • In one embodiment, the program codes for determining a covariance matrix of the N target images are executed by the processor 3000 for:
  • carrying out target segmentation for the N target images respectively to obtain the N face images; and
  • determining the covariance matrix of the N face image.
  • In one embodiment, the program codes for acquiring N target images are executed by the processor 3000 for:
  • acquiring M first images, wherein M is a positive integer; and
  • performing image segmentation on the M first images to obtain the N target images.
  • In one embodiment, the program codes for acquiring N target images are executed by the processor 3000 for:
  • acquiring N second images;
  • determining an image quality evaluation value of each of the N second images to obtain N image quality evaluation values; and
  • performing image enhancement processing on the second images whose image quality evaluation values are less than a second pre-set threshold value, and keeping the second images whose image quality evaluation values are greater than or equal to the second pre-set threshold value to obtain the N target images.
  • In one embodiment, the N target images correspond to a target tag.
  • after the program codes are executed for determining an upper information limit according to the covariance matrix, the program codes are further executed by the processor 3000 for:
  • evaluating the accuracy of the target tag according to the upper information entropy limit.
  • In one embodiment, after the program codes are executed for evaluating the accuracy of the target tag according to the upper information entropy limit, the program codes are further executed by the processor 3000 for:
  • deleting the N target images when the accuracy is lower than a third pre-set threshold value; and
  • retaining the N target images when the accuracy is higher than or equal to the third pre-set threshold value.
  • The embodiment of the invention also provides a computer storage medium in which a program may be stored, and the execution of the program includes part or all steps of any of the image processing methods recorded in the embodiments.
  • Although the invention is described herein in combination with each embodiment above, other variations of the embodiments may be understood and realized by a person having ordinary skills in the field by viewing the figured, disclosed contents and the claims in the process of implementing the invention. In claims, the word “comprising” does not exclude the other steps of components, “a” or “an” does not exclude the multiple. A single processor or other unit may implement several functions enumerated in the claims. although some technical features are recorded in different dependent claims, this does not means that these technical features cannot be combined to produce good results.
  • A person having ordinary skills in the field can understand that embodiments of the invention may be provided as methods, devices (equipments), or computer program products. Therefore, the invention may take the form of full hardware embodiments, full software embodiments, or embodiments combining software and hardware aspects. Furthermore, the present invention may take the form of a computer program product implemented on one or more computer available storage media (including but not limited to magnetic disk storage, CD-ROM, optical memory, etc.) including computer available program codes. Computer programs are stored/distributed in suitable storage medium, provided with or as part of other hardware, and may also be distributed in other forms, such as the Internet or other wired or wireless telecommunications systems.
  • The present invention is described by reference to the method, device (equipment) and flow chart and/or block diagram of a computer program product of an embodiment of the present invention. Each flow and/or box in the flowchart and/or block diagram, and the combination of flow and/or box in the flowchart and/or block diagram, shall be understood to be implemented by computer program instructions. The computer program instructions can be provided to a general purpose computer, a special purpose computer, an embedded processor or processors of other programmable data processing devices to produce a machine, making the instructions executed by the processors of the computer or other programmable data processing devices to produce a device used to implement the designated function of a flow or multiple flow in a flow chart and/or a box or multiple box in a block diagram.
  • These computer program instructions can also be stored in the computer readable storage which can boot a computer or other programmable data processing equipment in a particular way to work, making the instructions stored in the computer readable storage produce manufactures including command devices, the command devices can implement the designated function of a flow or multiple flow in a flow chart and/or a box or multiple box in a block diagram.
  • The computer program instructions can be loaded into a computer or other programmable data processing equipment, making the computer or other programmable devices to perform a series of operating steps to produce the processing of computer implementation, therefore the instructions executed on a computer or other programmable devices can provide the steps for implementing the designated function of a flow or multiple flow in a flow chart and/or a box or multiple box in a block diagram.
  • Although the invention is described in combination with specific features and embodiments, it is evident that it can be modified and combined in various ways without departing from the spirit and scope of the invention. Accordingly, this specification and accompanying drawings are only exemplary descriptions of the invention as defined by the claims and are deemed to cover any and all modifications, variations, combinations or equivalents within the scope of the invention. The foregoing descriptions are merely exemplary embodiments of the present invention, but not intended to limit the protection scope of the present disclosure. Any variation or replacement made by persons of ordinary skills in the art without departing from the spirit of the present disclosure shall fall within the protection scope of the present disclosure. Therefore, the scope of the present disclosure shall be subject to be appended claims.

Claims (12)

1. An image processing method, comprising:
acquiring N target images, wherein each of the target images comprises a face image, and N is an integer greater than one;
determining a covariance matrix of the N target images;
determining an upper information entropy limit according to the covariance matrix;
determining whether the upper information entropy limit is greater than a first pre-set threshold value;
when the upper information entropy limit is greater than the first pre-set threshold value, determining that the N target images comprise face images of different people; and
when the upper information entropy limit is less than or equal to the first pre-set threshold value, determining that the N target images comprise face images of the same person.
2. The image processing method of claim 1, the step of determining a covariance matrix of the N target images, comprising:
extracting feature data of each target image in the N target images respectively to obtain N feature data;
calculating the covariance matrix according to the N feature data.
3. The image processing method of claim 1, the step of determining a covariance matrix of the N target images, comprising:
carrying out target segmentation for the N target images respectively to obtain the N face images;
determining the covariance matrix of the N face images.
4. The image processing method of claim 1, the step of acquiring N target images, comprising:
acquiring M first images, wherein M is a positive integer,
performing image segmentation on the M first images to obtain the N target images.
5. The image processing method of claim 1, the step of acquiring N target images, comprising:
acquiring N second images;
determining an image quality evaluation value of each of the N second images to obtain N image quality evaluation values;
performing image enhancement processing on the second images whose image quality evaluation values are less than a second pre-set threshold value, and keeping the second images whose image quality evaluation values are greater than or equal to the second pre-set threshold value to obtain the N target images.
6. The image processing method of claim 1, wherein the N target images correspond to a target tag;
after the step of determining an upper information entropy limit according to the covariance matrix, further comprising:
evaluating the accuracy of the target tag according to the upper information entropy limit.
7. The image processing method of claim 6, after the step of evaluating the accuracy of the target tag according to the upper information entropy limit, further comprising:
deleting the N target images when the accuracy is lower than a third pre-set threshold value;
retaining the N target images when the accuracy is higher than or equal to the third pre-set threshold value.
8. A terminal, comprising:
an acquiring unit, configured to acquire N target images, wherein each of the target images comprises a face image, and N is an integer greater than one;
a first determining unit, configured to determine a covariance matrix of the N target images;
a second determining unit, configured to determine an upper information entropy limit according to the covariance matrix;
a judging unit, configured to determine whether the upper information entropy limit is greater than a first pre-set threshold value;
a third determining unit, configured to determine that the N target images comprise face images of different people when the judging unit determines that the upper information entropy limit is greater than the first pre-set threshold value;
the third determining unit, further configured to determine that the N target images comprise face images of the same person when the judging unit determines that the upper information entropy limit is less than or equal to the first pre-set threshold value.
9. The terminal of claim 8, the first determining unit comprising:
an extracting module, configured to extract feature data of each target image in the N target images respectively to obtain N feature data;
a calculating module, configured to calculate the covariance matrix according to the N feature data.
10. The terminal of claim 8, the first determining unit comprising:
a first segmenting module, configured to segment the N target images respectively to obtain the N face images;
a first determining module configured to determine the covariance matrix of the N face image.
11. A terminal, comprising a processor, the processor configured to execute a computer program stored in a memory to implement:
acquiring N target images, wherein each of the target images comprises a face image, and N is an integer greater than one;
determining a covariance matrix of the N target images;
determining an upper information entropy limit according to the covariance matrix;
determining whether the upper information entropy limit is greater than a first pre-set threshold value;
when the upper information entropy limit is greater than the first pre-set threshold value, determining that the N target images comprise face images of different people; and
when the upper information entropy limit is less than or equal to the first pre-set threshold value, determining that the N target images comprise face images of the same person.
13. (canceled)
US16/461,789 2016-11-18 2017-11-15 Image processing method, terminal and storage medium Active US10467743B1 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
CN201611026182 2016-11-18
CN201611026182.5A CN106845331B (en) 2016-11-18 2016-11-18 A kind of image processing method and terminal
CN201611026182.5 2016-11-18
PCT/CN2017/111176 WO2018090937A1 (en) 2016-11-18 2017-11-15 Image processing method, terminal and storage medium

Publications (2)

Publication Number Publication Date
US10467743B1 US10467743B1 (en) 2019-11-05
US20190340744A1 true US20190340744A1 (en) 2019-11-07

Family

ID=59145774

Family Applications (1)

Application Number Title Priority Date Filing Date
US16/461,789 Active US10467743B1 (en) 2016-11-18 2017-11-15 Image processing method, terminal and storage medium

Country Status (3)

Country Link
US (1) US10467743B1 (en)
CN (1) CN106845331B (en)
WO (1) WO2018090937A1 (en)

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106845331B (en) * 2016-11-18 2018-03-13 深圳云天励飞技术有限公司 A kind of image processing method and terminal
CN108564102A (en) * 2018-01-04 2018-09-21 百度在线网络技术(北京)有限公司 Image clustering evaluation of result method and apparatus
CN110858286B (en) * 2018-08-23 2023-08-11 杭州海康威视数字技术股份有限公司 Image processing method and device for target recognition
CN109447006A (en) * 2018-11-01 2019-03-08 北京旷视科技有限公司 Image processing method, device, equipment and storage medium
CN112241748B (en) * 2019-07-16 2024-06-14 广州汽车集团股份有限公司 Data dimension reduction method and device based on multi-source information entropy difference
CN110991386B (en) * 2019-12-14 2023-04-18 长沙芯连心智慧***有限责任公司 Robust nearest neighbor radar target one-dimensional range profile identification method and device
CN111562854B (en) * 2020-04-08 2023-08-29 歌尔科技有限公司 Detection pad placement position selection method, device, equipment and storage medium
CN111696083B (en) * 2020-05-20 2024-05-14 平安科技(深圳)有限公司 Image processing method and device, electronic equipment and storage medium
CN111611417B (en) * 2020-06-02 2023-09-01 Oppo广东移动通信有限公司 Image de-duplication method, device, terminal equipment and storage medium
CN112526705A (en) * 2020-12-07 2021-03-19 中国科学院长春光学精密机械与物理研究所 On-orbit self-adaptive focusing method, device, equipment and medium for optical load of spacecraft
CN113506246B (en) * 2021-06-15 2022-11-25 西安建筑科技大学 Concrete 3D printing component fine detection method based on machine vision
CN113658229B (en) * 2021-08-13 2024-02-02 杭州华橙软件技术有限公司 Method and device for determining abnormal object, storage medium and electronic device

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101414348A (en) * 2007-10-19 2009-04-22 三星电子株式会社 Method and system for identifying human face in multiple angles
JP5376814B2 (en) * 2008-03-03 2013-12-25 三菱電機株式会社 Target identification device
TWI456409B (en) * 2011-10-18 2014-10-11 Univ Nat Sun Yat Sen Calculation method for ofdm system with papr reduction function
CN103605971B (en) * 2013-12-04 2017-05-24 深圳市捷顺科技实业股份有限公司 Method and device for capturing face images
CN103679158B (en) * 2013-12-31 2017-06-16 北京天诚盛业科技有限公司 Face authentication method and device
CN105138972B (en) * 2015-08-11 2020-05-19 北京眼神智能科技有限公司 Face authentication method and device
CN105550657B (en) * 2015-12-23 2019-01-29 北京化工大学 Improvement SIFT face feature extraction method based on key point
CN106845331B (en) 2016-11-18 2018-03-13 深圳云天励飞技术有限公司 A kind of image processing method and terminal

Also Published As

Publication number Publication date
US10467743B1 (en) 2019-11-05
CN106845331B (en) 2018-03-13
WO2018090937A1 (en) 2018-05-24
CN106845331A (en) 2017-06-13

Similar Documents

Publication Publication Date Title
US10467743B1 (en) Image processing method, terminal and storage medium
US20200042770A1 (en) Facial recognition method and apparatus, electronic device, and storage medium
US20190279014A1 (en) Method and apparatus for detecting object keypoint, and electronic device
US10891465B2 (en) Methods and apparatuses for searching for target person, devices, and media
Kaneko et al. Robust image registration by increment sign correlation
WO2021139324A1 (en) Image recognition method and apparatus, computer-readable storage medium and electronic device
CN109711508B (en) Image processing method and device
CN111625668B (en) Object detection and candidate filtering system
CN107209942B (en) Object detection method and image retrieval system
US8811671B2 (en) Image processing apparatus, image processing method, and recording medium
US10922535B2 (en) Method and device for identifying wrist, method for identifying gesture, electronic equipment and computer-readable storage medium
CN110910445B (en) Object size detection method, device, detection equipment and storage medium
CN109446061A (en) A kind of page detection method, computer readable storage medium and terminal device
CN109949295A (en) Otherness detection method, device and the computer storage medium of browsing device net page
US20230245339A1 (en) Method for Adjusting Three-Dimensional Pose, Electronic Device and Storage Medium
CN111444807B (en) Target detection method, device, electronic equipment and computer readable medium
CN112990318A (en) Continuous learning method, device, terminal and storage medium
US9081800B2 (en) Object detection via visual search
US20230096921A1 (en) Image recognition method and apparatus, electronic device and readable storage medium
CN115690672A (en) Abnormal image recognition method and device, computer equipment and storage medium
CN113378790B (en) Viewpoint positioning method, apparatus, electronic device, and computer-readable storage medium
CN112149570B (en) Multi-person living body detection method, device, electronic equipment and storage medium
Ghandour et al. Building shadow detection based on multi-thresholding segmentation
CN113762027B (en) Abnormal behavior identification method, device, equipment and storage medium
CN114663980A (en) Behavior recognition method, and deep learning model training method and device

Legal Events

Date Code Title Description
AS Assignment

Owner name: SHENZHEN INTELLIFUSION TECHNOLOGIES CO., LTD., CHI

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:WANG, DONGLU;REEL/FRAME:049205/0051

Effective date: 20190313

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4