CN109191426A - A kind of flat image conspicuousness detection method - Google Patents

A kind of flat image conspicuousness detection method Download PDF

Info

Publication number
CN109191426A
CN109191426A CN201810820639.2A CN201810820639A CN109191426A CN 109191426 A CN109191426 A CN 109191426A CN 201810820639 A CN201810820639 A CN 201810820639A CN 109191426 A CN109191426 A CN 109191426A
Authority
CN
China
Prior art keywords
layer
convolutional
path
network model
sampling
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201810820639.2A
Other languages
Chinese (zh)
Inventor
桑庆兵
殷莹
李朝锋
过榴晓
吴小俊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jiangnan University
Original Assignee
Jiangnan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jiangnan University filed Critical Jiangnan University
Priority to CN201810820639.2A priority Critical patent/CN109191426A/en
Publication of CN109191426A publication Critical patent/CN109191426A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Landscapes

  • Engineering & Computer Science (AREA)
  • Quality & Reliability (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

The present invention proposes a kind of flat image conspicuousness detection method, can export the higher notable figure of accuracy, and method is simpler, it can be readily appreciated that being suitable for the application of.It is the following steps are included: S1: component network model;S2: training network model, the network model after being trained;S3: image data is input to the network model after training obtained in S2, carries out conspicuousness detection, exports notable figure, is handled for subsequent image;The structure of network model includes the constricted path in a left side and the path expander on a right side in step sl;Constricted path is used to capture context comprising an input layer, multiple convolutional layers and pond layer;Convolution operation is carried out by incoming first convolutional layer of the tested altimetric image of input layer input;The continuous convolutional layer of every two followed by a pond layer;Path expander is used to be accurately positioned comprising multiple up-sampling layers and convolutional layer, an output layer;Each up-sampling layer followed by two continuous convolutional layers.

Description

A kind of flat image conspicuousness detection method
Technical field
The present invention relates to computer vision and digital image processing techniques field, specially a kind of flat image conspicuousness inspection Survey method.
Background technique
Conspicuousness detects a preprocessing process as image procossing, is widely used in image quality evaluation, target The fields such as tracking, target detection, image segmentation.Such as: there are two application of the conspicuousness detection in image quality evaluation is main Aspect calculates the salient region of image to be evaluated in feature extraction phases, a feature as image to be evaluated;? Fusion Features stage, the weight that the value of pixel in notable figure is merged as each characteristic pattern;Conspicuousness detection method is usually divided For two classes: from top to bottom with it is bottom-up, since bottom-up method does not need manual annotation as priori knowledge, so mesh Preceding research all concentrates in bottom-up method;In existing research, there are many method for conspicuousness detection, such as: Frequency-tuned salient region detection (hereinafter referred to as FT), Saliency detection via Graph-based manifold ranking (hereinafter referred to as GMR), Hierarchical saliency detection(with Lower abbreviation HS), however these algorithms are in a particular application, when conspicuousness object appears in image border or works as obvious object When similar with part background information, it is easily lost target information, it may appear that the inadequate problem of accuracy.
Summary of the invention
In order to solve the problems, such as that the notable figure accuracy exported in conspicuousness detection method is inadequate, the present invention proposes a kind of flat Face image significance detection method can export the higher notable figure of accuracy, and method is simpler, it can be readily appreciated that being suitable for answering With.
The technical scheme is that such: a kind of flat image conspicuousness detection method comprising following steps:
S1: component network model;
S2: the training network model, the network model after being trained;
S3: being input to the network model after the training obtained in S2 for image data, carries out conspicuousness detection, and output is significant Figure is handled for subsequent image;
It is characterized by: the structure of the network model includes constricted path and the right side on the left of in the of one in step sl Path expander;
The constricted path is used to capture context comprising an input layer, multiple convolutional layers and pond layer;By the input Incoming first convolutional layer of the tested altimetric image of layer input carries out convolution operation;Every two is continuously after the convolutional layer And then a pond layer, an and then activation primitive after each convolutional layer;
The path expander is used to be accurately positioned comprising multiple up-sampling layers and the convolutional layer, an output layer;Each institute State up-sampling layer followed by two continuous convolutional layers, an activation letter is followed after each convolutional layer Number;
The characteristic pattern obtained after the completion of the pondization operation of the pond layer of the last one in the constricted path, then input two institutes It states convolutional layer, carries out the continuous convolution operation twice, obtained characteristic pattern is then inputted the into the path expander One up-sampling layer carries out up-sampling operation;
The number of the up-sampling layer in the path expander is identical as the number of the pond layer in the constricted path;
The number of the convolutional layer in the path expander is identical as the number of the convolutional layer in the constricted path;
The characteristic pattern exported after the up-sampling operation in the up-sampling layer every time, needs and the contraction road at the same level Then the characteristic pattern connection of pond layer output in diameter inputs the subsequent convolutional layer in the path expander again and carries out The convolution operation;
A warp lamination before the output layer, the characteristic pattern that the warp lamination exports the previous convolutional layer into Row deconvolution operation, then cut, after normalization operation, pass through notable figure needed for output layer output.
It is further characterized by:
The convolution kernel of two convolutional layers before the warp lamination is sized to 1 × 1, remaining convolutional layer The size of convolution kernel is all set to 3 × 3;
The maximum pondization operation that the step-length that the pond layer is 2 × 2 is 2;
The activation primitive is Leaky Rectified Linear Units function, calculation formula are as follows:
A is constant;
Notable figure is normalized into [0,1] using Sigmoid function in the normalization operation;
In step s 2, in training network model process, to the notable figure of output layer output, pass through function Euclidean Loss carries out conspicuousness regression analysis.
Network model in technical solution provided by the invention, by constricted path and path expander group in present networks model At constricted path is for obtaining contextual information, and path expander is for accurately positioning, and two paths are symmetrical, forms U Shape;Notable figure is generated using warp lamination in the method, deconvolution operation is carried out to characteristic pattern, then cut to notable figure And normalization operation, then obtain final output is the interested object of human eye, and present networks model is made to be more suitable for conspicuousness survey Detection method;Because pondization operation can lose image information and reduce image resolution ratio and be irreversible operation, feature is influenced The final accuracy of figure, right half part is before doing convolution operation, by the resolution ratio of the pondization operation output of left-half peer Output characteristic pattern is connected together after relatively high characteristic pattern and up-sampling operation, high-resolution characteristic pattern and more abstract Characteristic pattern combines, and can supply the information of some pictures, to obtain more accurate notable figure.
Detailed description of the invention
Fig. 1 is network architecture schematic diagram of the invention.
Specific embodiment
It is by the invention that the technical scheme comprises the following steps:
S1: component network model;
S2: using THUS data set training network model, the network model after being trained;
S3: being input to the network model after training obtained in S2 for image data, carry out conspicuousness detection, export notable figure, For subsequent image processing;
As shown in Figure 1, the network model in S1 is based on Convolutional Networks for Biomedical Image Segmentation(hereinafter referred to as U-Net) network model and design, structure include one left side constricted path and one The path expander on a right side;Present networks model belongs to network end to end, and input is piece image, and output is also piece image.
In present networks figure, constricted path is used to capture context comprising an input layer, multiple convolutional layers and pond Change layer;Convolution operation is carried out by incoming first convolutional layer of the tested altimetric image of input layer input;The continuous convolutional layer of every two Followed by a pond layer, an and then activation primitive after each convolutional layer;The convolution of convolution operation in constricted path Core is set as 3 × 3;The maximum pondization operation that the step-length that pondization operation is 2 × 2 is 2.
Activation primitive is Leaky Rectified Linear Units (hereinafter referred to as L-ReLU) function, calculation formula Are as follows:
In L-ReLU function, when the input output of x>0 is x, as x<0, output is a*x, and a is constant, is taken in the present embodiment 0.1;Activation primitive in former U-net network uses Rectified Linear Unit(hereinafter referred to as ReLU) function, ReLU letter It will no longer be updated after several activation to big gradient, L-ReLU can solve this problem, so that be worth around AnchorPoint to retain, Activation primitive L-ReLU can greatly improve training speed, and can reduce the possibility of over-fitting, so that performance of the invention More preferably.
Pond layer, convolutional layer number be configured according to the actual situation before component network, this reality as shown in Figure 1 It applies and 4 pond layers and corresponding 4 up-samplings layer is set in example.
Path expander is used to be accurately positioned comprising multiple up-sampling layers and convolutional layer, an output layer;Each up-sampling Layer followed by two continuous convolutional layers, an and then activation primitive after each convolutional layer;
In constricted path after the completion of the pondization operation of the last one pond layer, then two convolutional layers are inputted, carried out continuous twice Then first up-sampling layer in obtained characteristic pattern input p-adic extension p path is carried out up-sampling operation by convolution operation;
The number of up-sampling layer in path expander is identical as the number of pond layer in constricted path;
The number of convolutional layer in path expander is identical as the number of the convolutional layer in constricted path;
The characteristic pattern exported after the up-sampling operation in up-sampling layer every time, exports with the pond layer in constricted path at the same level Characteristic pattern connection, then input subsequent convolutional layer in path expander again and carry out convolution operation;Pond layer is defeated in constricted path The relatively high characteristic pattern of rate respectively out is defeated after operating with up-sampling after replicating and cutting (Copy and Crop) Characteristic pattern out combines, and high-resolution characteristic pattern is combined with more abstract characteristic pattern, can supply the information of some pictures, It obtains positioning more accurate characteristic pattern;
Since the target of the technical solution of this method offer is to find the interested object of human eye from image, by original U- The last one convolutional layer of Net network is changed to warp lamination, i.e., is a warp lamination before output layer, warp lamination will be previous The characteristic pattern of a convolutional layer output carries out deconvolution operation, then is cut, normalized to notable figure using Sigmoid function After [0,1], pass through notable figure needed for output layer output;
The convolution kernel of two convolutional layers before warp lamination is sized to 1 × 1, the size of the convolution kernel of remaining convolutional layer It is all set to 3 × 3;
It is reduced by above-mentioned up-sampling operation and deconvolution operation, the size of characteristic pattern, makes itself and original tested altimetric image Size it is identical, make export image and original image it is closer, confirmation get up more intuitively.
In step s 2, THUS data set training pattern is used in training network model process, is exported to by output layer Notable figure, using function Euclidean loss carry out conspicuousness regression analysis, learn conspicuousness from data.
In order to test the effect of technical solution of the present invention, pass through two common data sets ASD and ECSSD and two kinds Evaluation metrics are assessed:
(1) Area Under roc Curve (hereinafter referred to as AUC) assesses obvious object detection performance;
(2) Mean absolute error (hereinafter referred to as MAE) indicates the notable figure and benchmark image of model output (Ground-truth) mean absolute error between;
For the value of AUC and MAE within the scope of 0-1, AUC value illustrates that performance is better closer to 1, and MAE value illustrates aobvious closer to 0 Work property is better.Final testing result is shown in table 1, as can be seen from the table, compares with traditional several methods, and the present invention proposes Method on ASD and ECSSD data set, performance and conspicuousness target are all very good, have and benchmark image (Ground- Truth) higher similarity.
Table 1 is performance indicator of each model on ASD and ECSSD data set.
It after technical solution of the present invention, can be trained using less image is end-to-end, with others nerve Network model is compared, and the pondization operation in network model of the invention in path expander is sampled operation substitution, is increased The resolution ratio of characteristic pattern, by the high-resolution features figure of constricted path in conjunction with the characteristic pattern of up-sampling operation output, based on knot Image information after conjunction, continuous convolution behaviour later, which can assemble, more accurately to be exported;By the result of assessment it is recognised that It is compared with traditional several methods, method proposed by the present invention has higher similar to benchmark image (Ground-truth) Degree.

Claims (6)

1. a kind of flat image conspicuousness detection method comprising following steps:
S1: component network model;
S2: the training network model, the network model after being trained;
S3: being input to the network model after the training obtained in S2 for image data, carries out conspicuousness detection, and output is significant Figure is handled for subsequent image;
It is characterized by: the structure of the network model includes constricted path and the right side on the left of in the of one in step sl Path expander;
The constricted path is used to capture context comprising an input layer, multiple convolutional layers and pond layer;By the input Incoming first convolutional layer of the tested altimetric image of layer input carries out convolution operation;Every two is continuously after the convolutional layer And then a pond layer, an and then activation primitive after each convolutional layer;
The path expander is used to be accurately positioned comprising multiple up-sampling layers and the convolutional layer, an output layer;Each institute State up-sampling layer followed by two continuous convolutional layers, an activation letter is followed after each convolutional layer Number;
The characteristic pattern obtained after the completion of the pondization operation of the pond layer of the last one in the constricted path, then input two institutes It states convolutional layer, carries out the continuous convolution operation twice, obtained characteristic pattern is then inputted the into the path expander One up-sampling layer carries out up-sampling operation;
The number of the up-sampling layer in the path expander is identical as the number of the pond layer in the constricted path;
The number of the convolutional layer in the path expander is identical as the number of the convolutional layer in the constricted path;
The characteristic pattern exported after the up-sampling operation in the up-sampling layer every time, needs and the contraction road at the same level Then the characteristic pattern connection of pond layer output in diameter inputs the subsequent convolutional layer in the path expander again and carries out The convolution operation;
A warp lamination before the output layer, the characteristic pattern that the warp lamination exports the previous convolutional layer into Row deconvolution operation, then cut, after normalization operation, pass through notable figure needed for output layer output.
2. a kind of flat image conspicuousness detection method according to claim 1, it is characterised in that: before the warp lamination The convolution kernels of two convolutional layers be sized to 1 × 1, the size of the convolution kernel of remaining convolutional layer is all arranged It is 3 × 3.
3. a kind of flat image conspicuousness detection method according to claim 1, it is characterised in that: the pond layer is 2 × 2 Step-length be 2 maximum pondization operation.
4. a kind of flat image conspicuousness detection method according to claim 1, it is characterised in that: the activation primitive is Leaky Rectified Linear Units function, calculation formula are as follows:
A is constant.
5. a kind of flat image conspicuousness detection method according to claim 1, it is characterised in that: in the normalization operation Notable figure is normalized into [0,1] using Sigmoid function.
6. a kind of flat image conspicuousness detection method according to claim 1, it is characterised in that: in step s 2, training During network model, to the notable figure of output layer output, conspicuousness recurrence is carried out by function Euclidean loss Analysis.
CN201810820639.2A 2018-07-24 2018-07-24 A kind of flat image conspicuousness detection method Pending CN109191426A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810820639.2A CN109191426A (en) 2018-07-24 2018-07-24 A kind of flat image conspicuousness detection method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810820639.2A CN109191426A (en) 2018-07-24 2018-07-24 A kind of flat image conspicuousness detection method

Publications (1)

Publication Number Publication Date
CN109191426A true CN109191426A (en) 2019-01-11

Family

ID=64936733

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810820639.2A Pending CN109191426A (en) 2018-07-24 2018-07-24 A kind of flat image conspicuousness detection method

Country Status (1)

Country Link
CN (1) CN109191426A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109635882A (en) * 2019-01-23 2019-04-16 福州大学 Salient object detection method based on multi-scale convolution feature extraction and fusion
CN110910364A (en) * 2019-11-16 2020-03-24 应急管理部沈阳消防研究所 Method for detecting electrical equipment easy to cause fire in three-section fire scene based on deep neural network
CN110946566A (en) * 2019-12-16 2020-04-03 成都天奥电子股份有限公司 Heart beat classification method, device, equipment and storage medium based on U-Net network
CN111914850A (en) * 2019-05-07 2020-11-10 百度在线网络技术(北京)有限公司 Picture feature extraction method, device, server and medium
CN113327304A (en) * 2021-05-28 2021-08-31 北京理工大学重庆创新中心 Hyperspectral image saliency map generation method based on end-to-end neural network

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106296692A (en) * 2016-08-11 2017-01-04 深圳市未来媒体技术研究院 Image significance detection method based on antagonism network
CN106909924A (en) * 2017-02-18 2017-06-30 北京工业大学 A kind of remote sensing image method for quickly retrieving based on depth conspicuousness
CN107229918A (en) * 2017-05-26 2017-10-03 西安电子科技大学 A kind of SAR image object detection method based on full convolutional neural networks
CN108171701A (en) * 2018-01-15 2018-06-15 复旦大学 Conspicuousness detection method based on U networks and confrontation study

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106296692A (en) * 2016-08-11 2017-01-04 深圳市未来媒体技术研究院 Image significance detection method based on antagonism network
CN106909924A (en) * 2017-02-18 2017-06-30 北京工业大学 A kind of remote sensing image method for quickly retrieving based on depth conspicuousness
CN107229918A (en) * 2017-05-26 2017-10-03 西安电子科技大学 A kind of SAR image object detection method based on full convolutional neural networks
CN108171701A (en) * 2018-01-15 2018-06-15 复旦大学 Conspicuousness detection method based on U networks and confrontation study

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
OLAF RONNEBERGER 等: "U-Net: Convolutional Networks for Biomedical Image Segmentation", 《NAVAB N., HORNEGGER J., WELLS W., FRANGI A. (EDS) MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION – MICCAI 2015. MICCAI 2015. LECTURE NOTES IN COMPUTER SCIENCE》 *
张顺 等: "深度卷积神经网络的发展及其在计算机视觉领域的应用", 《计算机学报》 *
毋立芳 等: "局部聚类分析的FCN-CNN云图分割方法", 《软件学报HTTP://KNS.CNKI.NET/KCMS/DETAIL/11.2560.TP.20171204.0646.010.HTML》 *
田娟秀 等: "医学图像分析深度学习方法研究与挑战", 《自动化学报》 *
肖朝霞 等: "图像语义分割问题研究综述", 《软件导刊》 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109635882A (en) * 2019-01-23 2019-04-16 福州大学 Salient object detection method based on multi-scale convolution feature extraction and fusion
CN109635882B (en) * 2019-01-23 2022-05-13 福州大学 Salient object detection method based on multi-scale convolution feature extraction and fusion
CN111914850A (en) * 2019-05-07 2020-11-10 百度在线网络技术(北京)有限公司 Picture feature extraction method, device, server and medium
CN111914850B (en) * 2019-05-07 2023-09-19 百度在线网络技术(北京)有限公司 Picture feature extraction method, device, server and medium
CN110910364A (en) * 2019-11-16 2020-03-24 应急管理部沈阳消防研究所 Method for detecting electrical equipment easy to cause fire in three-section fire scene based on deep neural network
CN110946566A (en) * 2019-12-16 2020-04-03 成都天奥电子股份有限公司 Heart beat classification method, device, equipment and storage medium based on U-Net network
CN110946566B (en) * 2019-12-16 2022-06-17 成都天奥电子股份有限公司 Heart beat classification method, device, equipment and storage medium based on U-Net network
CN113327304A (en) * 2021-05-28 2021-08-31 北京理工大学重庆创新中心 Hyperspectral image saliency map generation method based on end-to-end neural network

Similar Documents

Publication Publication Date Title
Liu et al. Multistage GAN for fabric defect detection
CN109191426A (en) A kind of flat image conspicuousness detection method
Raza et al. Mimo-net: A multi-input multi-output convolutional neural network for cell segmentation in fluorescence microscopy images
Liu et al. Leveraging unlabeled data for crowd counting by learning to rank
CN107437092B (en) The classification method of retina OCT image based on Three dimensional convolution neural network
Xie et al. Beyond classification: structured regression for robust cell detection using convolutional neural network
CN112132817B (en) Retina blood vessel segmentation method for fundus image based on mixed attention mechanism
CA3100642A1 (en) Multi-sample whole slide image processing in digital pathology via multi-resolution registration and machine learning
CN111368690B (en) Deep learning-based video image ship detection method and system under influence of sea waves
CN112884712B (en) Method and related device for classifying defects of display panel
CN109376767A (en) Retina OCT image classification method based on deep learning
CN110852396A (en) Sample data processing method for cervical image
CN113888412B (en) Image super-resolution reconstruction method for diabetic retinopathy classification
Thanikkal et al. Advanced plant leaf classification through image enhancement and canny edge detection
CN112420170B (en) Method for improving image classification accuracy of computer aided diagnosis system
CN113468996A (en) Camouflage object detection method based on edge refinement
CN109685830A (en) Method for tracking target, device and equipment and computer storage medium
Shete et al. TasselGAN: An application of the generative adversarial model for creating field-based maize tassel data
CN107564007A (en) The scene cut modification method and system of amalgamation of global information
CN111209873A (en) High-precision face key point positioning method and system based on deep learning
CN117392496A (en) Target detection method and system based on infrared and visible light image fusion
Li et al. Image segmentation based on improved unet
CN113537119B (en) Transmission line connecting part detection method based on improved Yolov4-tiny
CN108764287B (en) Target detection method and system based on deep learning and packet convolution
Liang et al. Image segmentation and recognition for multi-class chinese food

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20190111

RJ01 Rejection of invention patent application after publication