CN112215163B - Weighted post-processing method applied to face detection prediction frame - Google Patents

Weighted post-processing method applied to face detection prediction frame Download PDF

Info

Publication number
CN112215163B
CN112215163B CN202011092870.8A CN202011092870A CN112215163B CN 112215163 B CN112215163 B CN 112215163B CN 202011092870 A CN202011092870 A CN 202011092870A CN 112215163 B CN112215163 B CN 112215163B
Authority
CN
China
Prior art keywords
prediction
confidence
conf
weighted
prediction frame
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011092870.8A
Other languages
Chinese (zh)
Other versions
CN112215163A (en
Inventor
朱海明
瞿洪桂
孙家乐
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Sinonet Science and Technology Co Ltd
Original Assignee
Beijing Sinonet Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Sinonet Science and Technology Co Ltd filed Critical Beijing Sinonet Science and Technology Co Ltd
Priority to CN202011092870.8A priority Critical patent/CN112215163B/en
Publication of CN112215163A publication Critical patent/CN112215163A/en
Application granted granted Critical
Publication of CN112215163B publication Critical patent/CN112215163B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Probability & Statistics with Applications (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to the field of face detection, and discloses a weighted post-processing method applied to a face detection prediction frame, which comprises the steps of utilizing a face detector to carry out face detection on a picture to be detected; deleting the prediction box with the confidence coefficient lower than the first confidence coefficient threshold value; calculating generalized intersection ratios between the prediction frame with the highest confidence coefficient and the updated prediction frames; screening n prediction frames with the generalized intersection ratio larger than an intersection ratio threshold value from the updated prediction frames; obtaining a weighted prediction frame corresponding to the prediction frame with the highest confidence coefficient; deleting n prediction boxes; and obtaining a weighted post-processing result of the face detector. The invention utilizes more effective prediction data of the face detector, so that the position information and the confidence information of the deleted prediction frame are effectively utilized at the same time, the position of the output prediction frame is better corrected, the generalized cross-over ratio is introduced, and the final output position precision of the face detector is improved.

Description

Weighted post-processing method applied to face detection prediction frame
Technical Field
The invention relates to the field of face detection, in particular to a weighted post-processing method applied to a face detection prediction frame.
Background
With the wide application of the deep learning technology, the technologies of face recognition, face beautification and the like are more and more mature, and face detection is the first step in the application scenes, and the precision of the face detection has a great influence on the precision of other follow-up algorithms. In a face detection algorithm, a model predicts a plurality of coordinates which may be face frames and corresponding confidence levels, and the prediction frames cannot be used as final output, so that repeated prediction frames and prediction frames with low confidence levels need to be filtered out.
In the existing methods, a non-maximum value suppression (NMS) method is mostly adopted, and an intersection ratio (IoU) between a prediction frame with the highest confidence degree in model prediction frames and other prediction frames is calculated, the prediction frame with IoU being larger than a set threshold value is directly deleted, IoU calculated by the prediction frame with the highest confidence degree and other prediction frames is selected, the prediction frame with the highest confidence degree being larger than the set threshold value is deleted IoU again, and the steps are repeated until all the prediction frames are processed. In the calculation mode, only the prediction frame with the highest confidence coefficient of the current picture is considered, and other prediction frames which are calculated by IoU and have higher coincidence degree with the prediction frame are deleted, and the deleted prediction frames also contain the position information of the human face, so that more effective data are lost. IoU cannot distinguish different alignment modes between two prediction frames when used as a calculation mode of coincidence degree, and cannot accurately reflect the coincidence degree of the two prediction frames. In the final output result of the method, only one piece of face uses the information of one prediction frame, more effective data is not utilized, and the position precision of the finally output face frame is not high.
Disclosure of Invention
The invention provides a weighted post-processing method applied to a face detection prediction frame, thereby solving the problems in the prior art.
A weighting post-processing method applied to a face detection prediction frame comprises the following steps:
s1) collecting a picture to be detected, and performing face detection on the picture to be detected by using a face detector to obtain all prediction frames corresponding to the picture to be detected, wherein each prediction frame corresponds to a confidence coefficient;
s2), setting a first confidence threshold conf _ thresh1, deleting the prediction boxes with confidence lower than the first confidence threshold conf _ thresh1 from all the prediction boxes, and obtaining a plurality of prediction boxes with confidence higher than or equal to the first confidence threshold conf _ thresh 1;
s3) updating a plurality of prediction frames with confidence degrees higher than or equal to a first confidence degree threshold conf _ thresh1 to obtain a plurality of updated prediction frames, sequencing the plurality of updated prediction frames according to the sequence of the confidence degrees from high to low to obtain the prediction frame with the highest current confidence degree in the plurality of updated prediction frames, wherein the confidence degree of the prediction frame with the highest current confidence degree is Max _ conf, and calculating the generalized intersection and ratio between the prediction frame with the highest current confidence degree and the plurality of updated prediction frames;
s4) setting an intersection-to-parallel ratio threshold value iou _ thresh, screening n prediction frames with the generalized intersection-to-parallel ratio being larger than the intersection-to-parallel ratio threshold value iou _ thresh from the updated prediction frames, and obtaining position information and confidence degrees of the n prediction frames;
s5) obtaining a weighted prediction frame corresponding to the prediction frame with the highest confidence coefficient according to the position information and the confidence coefficient of the n prediction frames, and storing the information of the weighted prediction frame, wherein the information of the weighted prediction frame comprises the position information and the confidence coefficient of the weighted prediction frame;
s6) deleting n prediction boxes from among the prediction boxes having the confidence level higher than or equal to the first confidence level threshold conf _ thresh 1;
s7), determining whether all the prediction frames with confidence higher than or equal to the first confidence threshold conf _ thresh1 have been deleted, if yes, proceeding to step S8); if not, returning to the step S3);
s8), obtaining a plurality of weighted prediction frames, setting a second confidence threshold conf _ thresh2, and obtaining a weighted post-processing result of the face detector according to the second confidence threshold conf _ thresh 2.
In step S7), if the confidence levels of the prediction frames higher than or equal to the first confidence level threshold conf _ thresh1 are not all deleted, the process returns to step S3), the prediction frames with the confidence levels higher than or equal to the first confidence level threshold conf _ thresh1 are updated, that is, the prediction frames with the confidence levels higher than or equal to the first confidence level threshold conf _ thresh1 after deleting the n prediction frames are used as the updated prediction frames, the process loops from step S3) to step S7), the prediction frames with the confidence levels higher than or equal to the first confidence level threshold conf _ thresh1 are continuously updated, a part of the prediction frames are deleted while obtaining the weighted prediction frames, and finally the prediction frames with the confidence levels higher than or equal to the first confidence level threshold conf _ thresh1 are all deleted while obtaining the weighted prediction frames.
Further, in step S3), calculating generalized intersection ratios between the prediction boxes with the highest current confidence and the updated prediction boxes respectively, including the following steps:
s31) recording the prediction box with the highest current confidence coefficient as A, and recording the ith prediction box in the updated prediction boxes as Bi
S32) calculating the prediction box A and the ith prediction box B with the highest current confidenceiThe intersection area A ≈ B between themi
S33) calculating the prediction box A and the ith prediction box B with the highest current confidenceiArea of union between A and Bi
S34) according to the intersection area A ≈ BiAnd union area Ab-iObtaining the prediction frame A and the ith prediction frame B with the highest current confidenceiGeneralized cross-over ratio between
Figure BDA0002722744550000031
C represents the prediction frame A with the highest current confidence coefficient and the ith prediction frame BiThe minimum envelope rectangular area.
Further, in step S5), a weighted prediction frame corresponding to the prediction frame with the highest current confidence is obtained according to the position information and the confidence of the n prediction frames, and information of the weighted prediction frame is stored, where the information of the weighted prediction frame includes the position information and the confidence of the weighted prediction frame, and the method includes the following steps:
s51) position information of the n prediction frames
Figure BDA0002722744550000041
Correspondingly multiplying the confidence degrees of the n prediction frames to obtain a position information matrix of the n prediction frames multiplied by the confidence degrees
Figure BDA0002722744550000042
The position information of the nth prediction box is [ X1 ]nY1n X2n Y2n],X1n、Y1nX coordinate value of the top left corner and y coordinate value of the top left corner of the nth prediction box, X2n、Y2nThe coordinate values of the lower right corner x and the lower right corner y of the nth prediction frame are respectively; confnConfidence for the nth prediction box;
s52) adding the position information matrixes of the n prediction frames multiplied by the confidence level in the step S51) row by row to obtain the position information of the prediction frames added row by row[X1 Y1 X2 Y2]X1 ═ conf, the coordinate value of the top left corner X of the prediction frame added row by row1*X11+conf2*X12+...+confn*X1nY1 (conf) is the coordinate value of the top left corner of the prediction frame added row by row1*Y11+conf2*Y12+...+confn*Y1nThe X-coordinate value X2 ═ conf at the bottom right corner of the prediction frame after line addition1*X21+conf2*X22+...+confn*X2nThe Y-coordinate value Y2 ═ conf at the bottom right corner of the prediction frame after line addition1*Y21+conf2*Y22+...+confn*Y2n
S53) calculating confidence sums conf ═ conf of the n prediction frames1+conf2+...+confn
S54) obtaining the weighted prediction frame corresponding to the prediction frame with the highest current confidence coefficient, and storing the information of the weighted prediction frame
Figure BDA0002722744550000043
Taking the confidence coefficient Max _ conf of the prediction frame with the highest current confidence coefficient as the confidence coefficient of the weighted prediction frame corresponding to the prediction frame with the highest current confidence coefficient, wherein the information of the weighted prediction frame comprises the position information of the weighted prediction frame
Figure BDA0002722744550000044
And the confidence Max _ conf of the weighted prediction box.
Further, in step S8), the weighted prediction result of the face detector is the position information and the confidence of several weighted prediction frames with confidence degrees greater than the second confidence degree threshold conf _ thresh2 and several weighted prediction frames with confidence degrees greater than the second confidence degree threshold conf _ thresh 2.
The invention has the beneficial effects that: compared with the NMS algorithm, the method utilizes more effective prediction data of the face detector, uses all the weight of the prediction frames with the highest confidence coefficient and the coincidence degree of the prediction frames with the highest confidence coefficient higher than a certain confidence coefficient threshold value as the final output (namely the weighted prediction frame) corresponding to the prediction frame with the highest confidence coefficient, and does not only use one prediction frame with the highest confidence coefficient and delete other coincident prediction frames, so that the position information of the deleted prediction frames is effectively utilized; when the weight processing of the prediction frames is carried out at the same time, the position information of each prediction frame is multiplied by the confidence coefficient of the prediction frame, the confidence coefficient information of the prediction frames is effectively utilized, and the output position of the prediction frame is better corrected; and a generalized intersection ratio GIoU which can more accurately evaluate the contact ratio of the two prediction frames is introduced, the generalized intersection ratio GIoU is used as a calculation mode of the contact ratio of the prediction frame with the highest confidence coefficient and other prediction frames, the selected data is more accurate, and the position accuracy finally output by the face detector is further improved.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings used in the embodiments are briefly described below, and it is obvious that the drawings in the following description are some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings without creative efforts.
Fig. 1 is a schematic flow chart of a weighting post-processing method applied to a face detection prediction frame according to a first embodiment of the present invention.
Fig. 2 is a schematic diagram of a minimum bounding rectangle area provided in the first embodiment.
Fig. 3 is a test result diagram comparing with other face detection algorithms according to the first embodiment.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is further described in detail below with reference to the accompanying drawings. It should be understood that the detailed description and specific examples, while indicating the invention, are intended for purposes of illustration only and are not intended to limit the scope of the invention.
In a first embodiment, a weighted post-processing method applied to a face detection prediction frame, as shown in fig. 1, includes the following steps:
s1) collecting a picture to be detected, and performing face detection on the picture to be detected by using a face detector to obtain all prediction frames corresponding to the picture to be detected, wherein each prediction frame corresponds to a confidence coefficient;
s2), setting a first confidence threshold conf _ thresh1, deleting the prediction boxes with confidence lower than the first confidence threshold conf _ thresh1 from all the prediction boxes, and obtaining a plurality of prediction boxes with confidence higher than or equal to the first confidence threshold conf _ thresh 1;
s3) updating a plurality of prediction frames with confidence degrees higher than or equal to a first confidence degree threshold conf _ thresh1, obtaining a plurality of updated prediction frames, sequencing the plurality of updated prediction frames according to the sequence of the confidence degrees from high to low, obtaining the prediction frame with the highest current confidence degree in the plurality of updated prediction frames, wherein the confidence degree of the prediction frame with the highest current confidence degree is Max _ conf, and calculating the generalized intersection and comparison between the prediction frame with the highest current confidence degree and the plurality of updated prediction frames respectively, wherein the method comprises the following steps:
s31) recording the prediction box with the highest current confidence coefficient as A, and recording the ith prediction box in other prediction boxes as Bi
S32) calculating the prediction box A and the ith prediction box B with the highest current confidenceiThe intersection area A ≈ B between themi
S33) calculating the prediction box A and the ith prediction box B with the highest current confidenceiArea of union between A and Bi
S34) according to the intersection area A ≈ BiAnd union area Ab-iObtaining the prediction frame A and the ith prediction frame B with the highest current confidenceiGeneralized cross-over ratio between
Figure BDA0002722744550000061
C represents the prediction frame A with the highest current confidence coefficient and the ith prediction frame BiThe minimum envelope rectangular area (see fig. 2).
S4) setting an intersection-to-parallel ratio threshold value iou _ thresh, screening n prediction frames with the generalized intersection-to-parallel ratio being larger than the intersection-to-parallel ratio threshold value iou _ thresh from the updated prediction frames, and obtaining position information and confidence degrees of the n prediction frames;
s5) obtaining a weighted prediction frame corresponding to the prediction frame with the highest current confidence according to the position information and the confidence of the n prediction frames, and storing information of the weighted prediction frame, where the information of the weighted prediction frame includes the position information and the confidence of the weighted prediction frame, and the method includes the following steps:
s51) position information of the n prediction frames
Figure BDA0002722744550000071
Correspondingly multiplying the confidence degrees of the n prediction frames to obtain a position information matrix of the n prediction frames multiplied by the confidence degrees
Figure BDA0002722744550000072
The position information of the nth prediction box is [ X1 ]nY1n X2n Y2n],X1n、Y1nX coordinate value of the top left corner and y coordinate value of the top left corner of the nth prediction box, X2n、Y2nThe coordinate values of the lower right corner x and the lower right corner y of the nth prediction frame are respectively; confnConfidence for the nth prediction box;
s52) adding the position information matrixes of the n prediction frames multiplied by the confidence degrees in S51) row by row to obtain the position information [ X1Y 1X 2Y 2 of the prediction frames added row by row]X1 ═ conf, the coordinate value of the top left corner X of the prediction frame added row by row1*X11+conf2*X12+...+confn*X1nY1 (conf) is the coordinate value of the top left corner of the prediction frame added row by row1*Y11+conf2*Y12+...+confn*Y1nThe X-coordinate value X2 ═ conf at the bottom right corner of the prediction frame after line addition1*X21+conf2*X22+...+confn*X2nThe Y-coordinate value Y2 ═ conf at the bottom right corner of the prediction frame after line addition1*Y21+conf2*Y22+...+confn*Y2n
S53) calculating confidence sums conf ═ conf of the n prediction frames1+conf2+...+confn
S54) obtaining the weighted prediction frame corresponding to the prediction frame with the highest current confidence coefficient, and saving the information of the weighted prediction frame
Figure BDA0002722744550000081
Taking the confidence coefficient Max _ conf of the prediction frame with the highest current confidence coefficient as the confidence coefficient of the weighted prediction frame corresponding to the prediction frame with the highest current confidence coefficient, wherein the information of the weighted prediction frame comprises the position information of the weighted prediction frame
Figure BDA0002722744550000082
And the confidence Max _ conf of the weighted prediction box.
S6) deleting n prediction boxes from among the prediction boxes having the confidence level higher than or equal to the first confidence level threshold conf _ thresh 1;
s7), determining whether all the prediction frames with confidence higher than or equal to the first confidence threshold conf _ thresh1 have been deleted, if yes, proceeding to step S8); if not, returning to the step S3);
s8), obtaining a plurality of weighted prediction frames, setting a second confidence threshold conf _ thresh2, and obtaining a weighted post-processing result of the face detector according to the second confidence threshold conf _ thresh 2.
In step S8), the weighted post-processing results of the face detector are position information and confidence of several weighted prediction frames with confidence degrees greater than the second confidence threshold conf _ thresh2 and several weighted prediction frames with confidence degrees greater than the second confidence threshold conf _ thresh 2.
Compared with other face detection algorithms, the technical scheme adopted by the embodiment is optimal in accuracy, less increase in running time is caused, and specific test results are shown in fig. 3. In FIG. 3, other various face detection algorithms include NMS algorithm, soft-NMS algorithm, and-NMS (GIoU) algorithm, and-NMS (DIoU) algorithm, weighted-NMS (GIoU) algorithm, weighted-NMS (DIoU) algorithm, DIoU-NMS algorithm, and GIoU-NMS algorithm; taking the intersection ratio threshold value iou _ thresh of other face detection algorithms except the soft-NMS algorithm to be 0.5; confidence threshold 2 is a second confidence threshold conf _ thresh 2; the AP (average precision) reflects the accuracy of the prediction box, with greater numbers being more accurate. The technical scheme adopted by the invention is a weighted-NMS (GIoU) algorithm, and it can be seen that the APs of the technical scheme adopted by the invention (namely the weighted-NMS (GIoU) algorithm) are the highest under different confidence thresholds, and the improvement is the largest compared with the NMS algorithm.
By adopting the technical scheme disclosed by the invention, the following beneficial effects are obtained:
compared with the NMS algorithm, the method utilizes more effective prediction data of the face detector, uses all the weight of the prediction frames with the highest confidence coefficient and the coincidence degree of the prediction frames with the highest confidence coefficient higher than a certain confidence coefficient threshold value as the final output (namely the weighted prediction frame) corresponding to the prediction frame with the highest confidence coefficient, and does not only use one prediction frame with the highest confidence coefficient and delete other coincident prediction frames, so that the position information of the deleted prediction frames is effectively utilized; when the weight processing of the prediction frames is carried out at the same time, the position information of each prediction frame is multiplied by the confidence coefficient of the prediction frame, the confidence coefficient information of the prediction frames is effectively utilized, and the output position of the prediction frame is better corrected; and a generalized intersection ratio GIoU which can more accurately evaluate the contact ratio of the two prediction frames is introduced, the generalized intersection ratio GIoU is used as a calculation mode of the contact ratio of the prediction frame with the highest confidence coefficient and other prediction frames, the selected data is more accurate, and the position accuracy finally output by the face detector is further improved.
The foregoing is only a preferred embodiment of the present invention, and it should be noted that it will be apparent to those skilled in the art that various modifications and improvements can be made without departing from the principle of the present invention, and such modifications and improvements should also be considered within the scope of the present invention.

Claims (3)

1. A weighted post-processing method applied to a face detection prediction frame is characterized by comprising the following steps:
s1) collecting a picture to be detected, and performing face detection on the picture to be detected by using a face detector to obtain all prediction frames corresponding to the picture to be detected, wherein each prediction frame corresponds to a confidence coefficient;
s2), setting a first confidence threshold conf _ thresh1, deleting the prediction boxes with confidence lower than the first confidence threshold conf _ thresh1 from all the prediction boxes, and obtaining a plurality of prediction boxes with confidence higher than or equal to the first confidence threshold conf _ thresh 1;
s3) updating a plurality of prediction frames with confidence degrees higher than or equal to the first confidence degree threshold conf _ thresh1 to obtain a plurality of updated prediction frames, sequencing the plurality of updated prediction frames according to the sequence of the confidence degrees from high to low to obtain the prediction frame with the highest current confidence degree in the plurality of updated prediction frames, wherein the confidence degree of the prediction frame with the highest current confidence degree is Max _ conf, and calculating the generalized intersection-parallel ratio between the prediction frame with the highest current confidence degree and the plurality of updated prediction frames;
s4) setting an intersection-to-parallel ratio threshold value iou _ thresh, screening n prediction frames with the generalized intersection-to-parallel ratio being larger than the intersection-to-parallel ratio threshold value iou _ thresh from the updated prediction frames, and acquiring the position information and the confidence coefficient of the n prediction frames;
s5) obtaining a weighted prediction frame corresponding to the prediction frame with the highest confidence coefficient according to the position information and the confidence coefficient of the n prediction frames, and storing the information of the weighted prediction frame, wherein the information of the weighted prediction frame comprises the position information of the weighted prediction frame and the confidence coefficient of the weighted prediction frame; the method comprises the following steps:
s51) the position information of the n prediction frames
Figure FDA0003011011900000011
Correspondingly multiplying the confidence degrees of the n prediction frames to obtain a position information matrix of the n prediction frames multiplied by the confidence degrees
Figure FDA0003011011900000021
The position information of the nth prediction box is [ X1 ]n Y1nX2n Y2n],X1n、Y1nAre respectively asThe X-coordinate value of the upper left corner and the y-coordinate value of the upper left corner of the nth prediction box, X2n、Y2nThe coordinate values of the lower right corner x and the lower right corner y of the nth prediction frame are respectively; confnConfidence for the nth prediction box;
s52) adding the position information matrixes of the n prediction frames multiplied by the confidence level in the step S51) row by row to obtain the position information [ X1Y 1X 2Y 2 of the prediction frames added row by row]X1 ═ conf, the coordinate value of the top left corner X of the prediction frame added row by row1*X11+conf2*X12+...+confn*X1nY1 (conf) is the coordinate value of the top left corner of the prediction frame added row by row1*Y11+conf2*Y12+...+confn*Y1nThe X-coordinate value X2 ═ conf at the bottom right corner of the prediction frame after line addition1*X21+conf2*X22+...+confn*X2nThe Y-coordinate value Y2 ═ conf at the bottom right corner of the prediction frame after line addition1*Y21+conf2*Y22+...+confn*Y2n
S53) calculating confidence sums conf ═ conf of the n prediction frames1+conf2+...+confn
S54) obtaining the weighted prediction frame corresponding to the prediction frame with the highest current confidence coefficient, and saving the information of the weighted prediction frame
Figure FDA0003011011900000022
Taking the confidence coefficient Max _ conf of the prediction frame with the highest current confidence coefficient as the confidence coefficient of the weighted prediction frame corresponding to the prediction frame with the highest current confidence coefficient, wherein the information of the weighted prediction frame comprises the position information of the weighted prediction frame
Figure FDA0003011011900000023
And a confidence Max _ conf of the weighted prediction box;
s6) deleting the n prediction boxes from among the number of prediction boxes whose confidence is higher than or equal to the first confidence threshold conf _ thresh 1;
s7), determining whether all the prediction frames with confidence higher than or equal to the first confidence threshold conf _ thresh1 have been deleted, if yes, proceeding to step S8); if not, returning to the step S3);
s8), obtaining a plurality of weighted prediction frames, setting a second confidence coefficient threshold conf _ thresh2, and obtaining a weighted post-processing result of the face detector according to the second confidence coefficient threshold conf _ thresh 2.
2. The method for weighted post-processing of a face detection prediction box according to claim 1, wherein in step S3), the method for calculating the generalized intersection-parallel ratio between the prediction box with the highest current confidence coefficient and the updated prediction boxes comprises the following steps:
s31) recording the prediction box with the highest current confidence coefficient as A, and recording the ith prediction box in the updated prediction boxes as Bi
S32) calculating the prediction box A and the ith prediction box B with the highest current confidenceiThe intersection area between the two is A and Bi
S33) calculating the prediction box A and the ith prediction box B with the highest current confidenceiArea of union between A and Bi
S34) according to the intersection area A ^ BiAnd union area Ab-iObtaining the prediction frame A and the ith prediction frame B with the highest current confidenceiGeneralized cross-over ratio between
Figure FDA0003011011900000031
C represents the prediction frame A with the highest current confidence coefficient and the ith prediction frame BiThe minimum envelope rectangular area.
3. The method for weighted post-processing applied to a face detection prediction box according to claim 1, wherein in step S8), the result of weighted post-processing of the face detector is position information and confidence of several weighted prediction boxes with confidence degrees greater than the second confidence threshold conf _ thresh2 and several weighted prediction boxes with confidence degrees greater than the second confidence threshold conf _ thresh 2.
CN202011092870.8A 2020-10-13 2020-10-13 Weighted post-processing method applied to face detection prediction frame Active CN112215163B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011092870.8A CN112215163B (en) 2020-10-13 2020-10-13 Weighted post-processing method applied to face detection prediction frame

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011092870.8A CN112215163B (en) 2020-10-13 2020-10-13 Weighted post-processing method applied to face detection prediction frame

Publications (2)

Publication Number Publication Date
CN112215163A CN112215163A (en) 2021-01-12
CN112215163B true CN112215163B (en) 2021-05-25

Family

ID=74053965

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011092870.8A Active CN112215163B (en) 2020-10-13 2020-10-13 Weighted post-processing method applied to face detection prediction frame

Country Status (1)

Country Link
CN (1) CN112215163B (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107679469A (en) * 2017-09-22 2018-02-09 东南大学—无锡集成电路技术研究所 A kind of non-maxima suppression method based on deep learning
CN109886128A (en) * 2019-01-24 2019-06-14 南京航空航天大学 A kind of method for detecting human face under low resolution
CN110580445A (en) * 2019-07-12 2019-12-17 西北工业大学 Face key point detection method based on GIoU and weighted NMS improvement
CN110688994A (en) * 2019-12-10 2020-01-14 南京甄视智能科技有限公司 Human face detection method and device based on cross-over ratio and multi-model fusion and computer readable storage medium
CN111626990A (en) * 2020-05-06 2020-09-04 北京字节跳动网络技术有限公司 Target detection frame processing method and device and electronic equipment

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8824808B2 (en) * 2011-08-19 2014-09-02 Adobe Systems Incorporated Methods and apparatus for automated facial feature localization
CN106327469B (en) * 2015-06-29 2019-06-18 北京航空航天大学 A kind of video picture segmentation method of semantic label guidance
US9875429B2 (en) * 2015-10-06 2018-01-23 Adobe Systems Incorporated Font attributes for font recognition and similarity
CN111091551A (en) * 2019-12-12 2020-05-01 哈尔滨市科佳通用机电股份有限公司 Method for detecting loss fault of brake beam strut opening pin of railway wagon
CN111476817A (en) * 2020-02-27 2020-07-31 浙江工业大学 Multi-target pedestrian detection tracking method based on yolov3

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107679469A (en) * 2017-09-22 2018-02-09 东南大学—无锡集成电路技术研究所 A kind of non-maxima suppression method based on deep learning
CN109886128A (en) * 2019-01-24 2019-06-14 南京航空航天大学 A kind of method for detecting human face under low resolution
CN110580445A (en) * 2019-07-12 2019-12-17 西北工业大学 Face key point detection method based on GIoU and weighted NMS improvement
CN110688994A (en) * 2019-12-10 2020-01-14 南京甄视智能科技有限公司 Human face detection method and device based on cross-over ratio and multi-model fusion and computer readable storage medium
CN111626990A (en) * 2020-05-06 2020-09-04 北京字节跳动网络技术有限公司 Target detection frame processing method and device and electronic equipment

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Improving Object DetectionWith One Line of Code;Navaneeth Bodla et al.;《arXiv》;20170808;第1-9页 *
雾霾天气车辆检测的改进YOLOv3 算法;黄开启 等;《重庆大学学报》;20200911;第1-9页 *

Also Published As

Publication number Publication date
CN112215163A (en) 2021-01-12

Similar Documents

Publication Publication Date Title
US11450146B2 (en) Gesture recognition method, apparatus, and device
US11263434B2 (en) Fast side-face interference resistant face detection method
US8755623B2 (en) Image enhancement method, image enhancement device, object detection method, and object detection device
WO2023082882A1 (en) Pose estimation-based pedestrian fall action recognition method and device
CN103093212B (en) The method and apparatus of facial image is intercepted based on Face detection and tracking
US20040213460A1 (en) Method of human figure contour outlining in images
US9262672B2 (en) Pattern recognition apparatus and pattern recognition method that reduce effects on recognition accuracy, and storage medium
CN108320306B (en) Video target tracking method fusing TLD and KCF
CN111553425B (en) Template matching LSP algorithm, medium and equipment for visual positioning
CN101470809A (en) Moving object detection method based on expansion mixed gauss model
US20020122593A1 (en) Pattern recognition method and apparatus
US12002254B2 (en) Method and apparatus of training object detection network and object detection method and apparatus
CN110188780B (en) Method and device for constructing deep learning model for positioning multi-target feature points
CN113205138B (en) Face and human body matching method, equipment and storage medium
CN109447006A (en) Image processing method, device, equipment and storage medium
CN110738688A (en) novel infrared ultra-weak moving target detection method
CN111178261A (en) Face detection acceleration method based on video coding technology
CN110349070B (en) Short video watermark detection method
US20110222759A1 (en) Information processing apparatus, information processing method, and program
CN113255648B (en) Sliding window frame selection method and terminal based on image recognition
CN112802104B (en) Loop detection method based on RGB-D camera
CN112215163B (en) Weighted post-processing method applied to face detection prediction frame
US20230368397A1 (en) Method and system for detecting moving object
CN110738082B (en) Method, device, equipment and medium for positioning key points of human face
CN112508168B (en) Frame regression neural network construction method based on automatic correction of prediction frame

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant