CN109919098A - The recognition methods of target object and device - Google Patents

The recognition methods of target object and device Download PDF

Info

Publication number
CN109919098A
CN109919098A CN201910176768.7A CN201910176768A CN109919098A CN 109919098 A CN109919098 A CN 109919098A CN 201910176768 A CN201910176768 A CN 201910176768A CN 109919098 A CN109919098 A CN 109919098A
Authority
CN
China
Prior art keywords
value
target object
image
obtains
level set
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910176768.7A
Other languages
Chinese (zh)
Other versions
CN109919098B (en
Inventor
童云飞
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangzhou Shiyuan Electronics Thecnology Co Ltd
Original Assignee
Guangzhou Shiyuan Electronics Thecnology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangzhou Shiyuan Electronics Thecnology Co Ltd filed Critical Guangzhou Shiyuan Electronics Thecnology Co Ltd
Priority to CN201910176768.7A priority Critical patent/CN109919098B/en
Publication of CN109919098A publication Critical patent/CN109919098A/en
Application granted granted Critical
Publication of CN109919098B publication Critical patent/CN109919098B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses a kind of recognition methods of target object and devices.Wherein, this method comprises: obtaining the first image comprising target object;Image segmentation is carried out to the first image using U-shaped full convolutional neural networks model, obtains the boundary of target object, wherein the full convolutional neural networks model of U-shaped uses level set loss function.The recognition accuracy that the present invention solves target object in the prior art is lower, and the technical problem that repeatability is poor.

Description

The recognition methods of target object and device
Technical field
The present invention relates to field of image recognition, recognition methods and device in particular to a kind of target object.
Background technique
In the image recognition of medical field, leading portion optical coherence tomography (Anterior segment optical Coherence tomography, AS-OCT) it is used to many ophthalmology diseases of auxiliary diagnosis, such as disease of cornea, cataract and blueness Light eye etc. is a kind of style of shooting of non-intrusion type fanout free region.Crystalline lens is the main refraction structure of eyeball, when crystalline lens goes out When existing turbid phenomenon (density increase), vision disorder can be caused, cataract is caused to generate.Crystalline volume density is to measure cataract etc. The important indicator of disease severity, crystalline lens are the structures of a multilayer, can be roughly divided into from outside to inside: vitreous layer, cortex Layer and core.It is all performed manually by mostly currently based on the lens structure segmentation of AS-OCT image, and has repeatability poor, The defects of cost of labor is higher.It is obscured between lens structure boundary in AS-OCT image, causes full-automatic dividing difficult, especially It is the boundary of core (nucleus) and cortex (cortex).Moreover, the AS-OCT image shot in cataract patient eye, boundary It can become more to obscure.
It is lower for the recognition accuracy of target object in the prior art, and the problem that repeatability is poor, at present not yet It puts forward effective solutions.
Summary of the invention
The embodiment of the invention provides a kind of recognition methods of target object and devices, at least to solve mesh in the prior art The recognition accuracy for marking object is lower, and the technical problem that repeatability is poor.
According to an aspect of an embodiment of the present invention, a kind of recognition methods of target object is provided, comprising: acquisition includes First image of target object;Image segmentation is carried out to the first image using U-shaped full convolutional neural networks model, obtains target pair The boundary of elephant, wherein the full convolutional neural networks model of U-shaped uses level set loss function.
Further, the full convolutional neural networks model of U-shaped includes: encoding model and decoded model, and encoding model includes: more A first network block, first network block include: sequentially connected at least two first convolutional layer, and the first convolutional layer uses modified line Property unit activating function and pondization operation, in the last one first convolutional layer and next first network block in first network block First the first convolutional layer connection;Decoded model includes: multiple second network blocks and output layer, first network block and the second net The quantity of network block is identical, and the second network block includes: sequentially connected cascading layers and at least two second convolutional layers, cascading layers with it is right The first network block answered is connected with the last one second convolutional layer in upper second network block, cascading layers and corresponding first Network block is connected using duplication with combined jump connection type, the last one in output layer and the last one second network block Second side output layer connection, output layer use level set loss function.
Further, pondization operation includes one of following: maximum pondization operation and the operation of mean value pondization.
Further, level set loss function is determined as follows: the result for determining input to output layer is first Level set, and determine that the result of output layer output is the second level set;Obtain the shape and second level set of first level collection Shape;The state of shape and the second level set based on first level collection, obtains level set loss function.
Further, the shape of first level collection is obtained, comprising: obtain the first function value of first level collection, Yi Ji The second function value of the true value level set of one image;Based on first function value and second function value, the shape of first level collection is obtained Shape.
Further, it is based on first function value and second function value, obtains the shape of first level collection, comprising: obtains the The difference of two functional values and first function value, obtains difference;The absolute value for obtaining difference, obtains absolute value;Obtain the flat of absolute value Side, obtains square value;The integral for obtaining square value, obtains the shape of first level collection.
Further, after the second function value for obtaining true value level set, the above method further include: turned based on level set Exchange the letters number handles second function value, the second function value that obtains that treated;To treated, second function value is returned One change processing, obtains normalized function value;Based on first function value and normalized function value, the shape of first level collection is obtained.
Further, the shape of the second level set is obtained, comprising: multiple probability values of input to output layer are obtained, and The corresponding true value of each probability value;Based on each probability value and corresponding true value, the product of each probability value is obtained;It obtains multiple The sum of products of probability value obtains the shape of the second level set.
Further, it is based on each probability value and corresponding true value, obtains the product of each probability value, comprising: is obtained every The logarithm of a probability value obtains the logarithm of each probability value;Obtain the logarithm of each probability value and multiplying for corresponding true value Product, obtains the product of each probability value.
Further, the state of shape and the second level set based on first level collection, obtains level set loss function, packet It includes: obtaining the shape of first level collection and the product of the first parameter, obtain the first product;Obtain the shape and of the second level set The product of two parameters obtains the second product;The length on the boundary of target object and the product of third parameter are obtained, third is obtained and multiplies Product;The integrated value of target object region and the product of the 4th parameter are obtained, obtains the 4th product, wherein integrated value is used for Characterize the inside and outside of linking objective object region;Obtain the first product, the second product, the 4th product of third sum of products The sum of, obtain level set loss function.
Further, image segmentation is being carried out to the first image using U-shaped full convolutional neural networks model, is obtaining target pair Before the boundary of elephant, the above method further include: the first image is handled using edge detection algorithm, obtains target object institute The second image in region;Image segmentation is carried out to the second image using U-shaped full convolutional neural networks model, obtains target object Boundary.
Further, the first image is handled using edge detection algorithm, obtains the of target object region Two images, comprising: the first image is handled using multistage edge detection algorithm, obtains the second image.
Further, the first image comprising target object is obtained, comprising: utilize leading portion means of optical coherence tomography Target object is scanned, the first image is obtained.
Further, target object is lens nucleus.
According to another aspect of an embodiment of the present invention, a kind of identification device of target object is additionally provided, comprising: obtain mould Block, for obtaining the image comprising target object;Image segmentation module, for utilizing the full convolutional neural networks model of U-shaped to figure As carrying out image segmentation, the boundary of target object is obtained, wherein the full convolutional neural networks model of U-shaped loses letter using level set Number.
According to another aspect of an embodiment of the present invention, a kind of storage medium is additionally provided, storage medium includes the journey of storage Sequence, wherein equipment where control storage medium executes the recognition methods of above-mentioned target object in program operation.
According to another aspect of an embodiment of the present invention, a kind of processor is additionally provided, processor is used to run program, In, program executes the recognition methods of above-mentioned target object when running.
In embodiments of the present invention, it after getting the first image comprising target object, can use using water The full convolutional neural networks model of U-shaped of flat collection loss function carries out image segmentation to the first image, obtains the boundary of target object, Namely obtain final segmentation result.It is easily noted that, by combining the full convolutional neural networks model of U-shaped and level set to calculate Method carries out lens structure segmentation, realizes based on the full-automatic lens structure segmentation of deep learning, has reached and effectively improved crystalline substance The accuracy of shape body segmentation of structures and the technical effect of repeatability, and then the identification for solving target object in the prior art is quasi- True rate is lower, and the technical problem that repeatability is poor.
Detailed description of the invention
The drawings described herein are used to provide a further understanding of the present invention, constitutes part of this application, this hair Bright illustrative embodiments and their description are used to explain the present invention, and are not constituted improper limitations of the present invention.In the accompanying drawings:
Fig. 1 is a kind of flow chart of the recognition methods of target object according to an embodiment of the present invention;
Fig. 2 a is a kind of schematic diagram of optional AS-OCT image according to an embodiment of the present invention;
Fig. 2 b is a kind of schematic diagram of optional target object region according to an embodiment of the present invention;
Fig. 2 c is the schematic diagram for the result that one kind according to an embodiment of the present invention is optionally divided;
Fig. 3 is a kind of schematic diagram of the network structure of optional U-net network according to an embodiment of the present invention;
Fig. 4 a is a kind of schematic diagram of optional true value level set according to an embodiment of the present invention;
Fig. 4 b is a kind of schematic diagram of optional true value label according to an embodiment of the present invention;
Fig. 4 c is the schematic diagram of the result after a kind of optional normalization according to an embodiment of the present invention;And
Fig. 5 is a kind of schematic diagram of the identification device of target object according to an embodiment of the present invention.
Specific embodiment
In order to enable those skilled in the art to better understand the solution of the present invention, below in conjunction in the embodiment of the present invention Attached drawing, technical scheme in the embodiment of the invention is clearly and completely described, it is clear that described embodiment is only The embodiment of a part of the invention, instead of all the embodiments.Based on the embodiments of the present invention, ordinary skill people The model that the present invention protects all should belong in member's every other embodiment obtained without making creative work It encloses.
It should be noted that description and claims of this specification and term " first " in above-mentioned attached drawing, " Two " etc. be to be used to distinguish similar objects, without being used to describe a particular order or precedence order.It should be understood that using in this way Data be interchangeable under appropriate circumstances, so as to the embodiment of the present invention described herein can in addition to illustrating herein or Sequence other than those of description is implemented.In addition, term " includes " and " having " and their any deformation, it is intended that cover Cover it is non-exclusive include, for example, the process, method, system, product or equipment for containing a series of steps or units are not necessarily limited to Step or unit those of is clearly listed, but may include be not clearly listed or for these process, methods, product Or other step or units that equipment is intrinsic.
Embodiment 1
According to embodiments of the present invention, a kind of embodiment of the recognition methods of target object is provided, it should be noted that The step of process of attached drawing illustrates can execute in a computer system such as a set of computer executable instructions, also, It, in some cases, can be to be different from shown in sequence execution herein although logical order is shown in flow charts The step of out or describing.
Fig. 1 is a kind of flow chart of the recognition methods of target object according to an embodiment of the present invention, as shown in Figure 1, the party Method includes the following steps:
Step S102 obtains the first image comprising target object.
Optionally, above-mentioned target object can be lens nucleus.
Specifically, the first above-mentioned image can be the eyeground AS-OCT figure, include Phakic in image as shown in Figure 2 a Structure, wherein white box region is crystalline body region.Due to being obscured between lens nucleus and the boundary of cortex, Ke Yitong It crosses and the boundary of lens nucleus and cortex is accurately divided, realize the accurate segmentation to lens structure, hence, it can be determined that Target object is lens nucleus, and target object region is as shown in Figure 2 b.
Step S104 carries out image segmentation to the first image using the full convolutional neural networks model of U-shaped, obtains target object Boundary, wherein the full convolutional neural networks model of U-shaped use level set loss function.
Specifically, it in medical image segmentation field, is quickly grown currently based on deep learning image segmentation algorithm, especially It is the full convolutional neural networks model of U-shaped, the i.e. appearance of U-net network, in medicine figures such as Lung neoplasm, tumour and blood vessel thickness segmentations As field achieves biggish development.Therefore, can by U-net network as in AS-OCT image lens structure it is initial The segmentation of profile, determines lens nucleus region, and final segmentation result is as shown in Figure 2 c.
It should be noted that carrying out image segmentation to the eyeground AS-OCT figure by U-net network, crystalline substance can be not only determined Shape body core region, can also determine lenticular cornea and cortical area.
Due to the irregularity boundary that U-net network is partitioned into, the especially segmentation of lens nucleus, and Level Set Method is in shape Shape, noise resolution ratio and block etc. and to obtain preferable performance under some constraints, therefore, can be damaged by e-learning level set Lose function, the final segmentation result of the zero level set representations finally obtained, that is, another horizontal set identifier target object region Boundary.
Above-described embodiment through the invention can use use after getting the first image comprising target object The U-shaped full convolutional neural networks model of level set loss function carries out image segmentation to the first image, obtains target object Boundary, namely obtain final segmentation result.It is easily noted that, by combining the full convolutional neural networks model of U-shaped and water Flat set algorithm carries out lens structure segmentation, realizes based on the full-automatic lens structure segmentation of deep learning, has reached effective The accuracy of lens structure segmentation and the technical effect of repeatability are improved, and then solves target object in the prior art Recognition accuracy is lower, and the technical problem that repeatability is poor.
Optionally, the full convolutional neural networks model of U-shaped includes: encoding model and decoded model, and encoding model includes: multiple First network block, first network block include: sequentially connected at least two first convolutional layer, and the first convolutional layer is linear using amendment Unit activating function and pondization operate, in the last one first convolutional layer and next first network block in first network block First the first convolutional layer connection;Decoded model includes: multiple second network blocks and output layer, first network block and the second network The quantity of block is identical, and the second network block includes: sequentially connected cascading layers and at least two second convolutional layers, cascading layers with it is corresponding First network block connected with the last one second convolutional layer in upper second network block, cascading layers and corresponding first net Network block is connected using duplication with combineds jump connection type, the last one in output layer and the last one second network block the Two side output layer connection, output layer use level set loss function.
Optionally, above-mentioned pondization operation may include one of following: maximum pondization operation and the operation of mean value pondization.
It specifically,, can in order to improve image segmentation accuracy and repeatability since the size of the eyeground AS-OCT figure is larger To improve to existing U-net network, improvement is essentially consisted in: the size of input picture is larger, for example, input having a size of 1024*1024, that is, the size of the first image is 1024*1024;The convolution kernel that convolutional layer uses is smaller, for example, it may be 3* 3;The level of entire U-net network is more, for example, the level of whole network is 6 layers, that is, first network block and the second network The quantity of block is 6.
In embodiments of the present invention, activation primitive (Relu) and maximum pondization operation can be used in each convolutional layer.
The network structure of U-net network as shown in figure 3, network structure include encoding model (as shown in left side in Fig. 3) and Decoded model (as shown in right side in Fig. 3), VGG19 (Visual Geometry Group Network) can be used as network Coded portion, including six network blocks (i.e. above-mentioned first network block), each network block includes two to three convolutional layers, each A convolutional layer can all use activation primitive Relu (amendment linear unit, rectified linear unit) and be used for down-sampling Step-length be 2 2*2 maximum pond, the convolution kernel of convolutional layer is 3*3, can be expressed as Conv<3X3>with Relu, have Conducive to the expression of minutia, test speed relative to deeper residual error network can effective guarantee, in each down-sampling In step, feature number of channels is all doubled;Decoder module also includes six network blocks (the second i.e. above-mentioned network block), each net Network block includes a cascading layers, from corresponding characteristic layer (i.e. above-mentioned corresponding first network block) and up-sampling (up- Sample) (i.e. above-mentioned corresponding second convolutional layer), coefficient 2 use two convolutional layers and side output layer side- later Output, convolutional layer carries out convolution algorithm using the convolution kernel of 3*3, and uses activation primitive (Relu).Encoding model and decoding Model can be used duplication and merge the jump connection type of (Copy and Merge), can efficiently use the spy of different layers Sign, ensures better segmentation effect;Last output layer can be using level set loss function as final loss.
Optionally, level set loss function is determined as follows: the result for determining input to output layer is the first water Flat collection, and determine that the result of output layer output is the second level set;Obtain the shape of first level collection and the shape of the second level set Shape;The state of shape and the second level set based on first level collection, obtains level set loss function.
Optionally, the shape of first level collection is obtained, comprising: obtain the first function value and first of first level collection The second function value of the true value level set of image;Based on first function value and second function value, the shape of first level collection is obtained.
Specifically, it is based on first function value and second function value, obtains the shape of first level collection, comprising: obtains second The difference of functional value and first function value, obtains difference;The absolute value for obtaining difference, obtains absolute value;Square of absolute value is obtained, Obtain square value;The integral for obtaining square value, obtains the shape of first level collection.
In addition, this method further includes following steps after the second function value for obtaining true value level set: being based on level set Transfer function handles second function value, the second function value that obtains that treated;To treated, second function value is carried out Normalized obtains normalized function value;Based on first function value and normalized function value, the shape of first level collection is obtained Shape.
Similarly, the shape of the second level set is obtained, comprising: obtain multiple probability values of input to output layer, and every The corresponding true value of a probability value;Based on each probability value and corresponding true value, the product of each probability value is obtained;It obtains multiple general The sum of products of rate value obtains the shape of the second level set.
Specifically, it is based on each probability value and corresponding true value, obtains the product of each probability value, comprising: is obtained each The logarithm of probability value obtains the logarithm of each probability value;The logarithm of each probability value and the product of corresponding true value are obtained, Obtain the product of each probability value.
Optionally, the state of shape and the second level set based on first level collection, obtains level set loss function, packet It includes: obtaining the shape of first level collection and the product of the first parameter, obtain the first product;Obtain the shape and of the second level set The product of two parameters obtains the second product;The length on the boundary of target object and the product of third parameter are obtained, third is obtained and multiplies Product;The integrated value of target object region and the product of the 4th parameter are obtained, obtains the 4th product, wherein integrated value is used for Characterize the inside and outside of linking objective object region;Obtain the first product, the second product, the 4th product of third sum of products The sum of, obtain level set loss function.
In a kind of optional scheme, level set φ is the image (the first i.e. above-mentioned image) of input, each pixel (x, Y) functional value of value namely level set is and surfaceMinimum Eustachian distance, w be segmentation target object where area Domain.It can be using the result of input to output layer as first level collection φpre, that is, in the training process, U-net network it is defeated It is out level set φpre, and using output layer output as a result, namely image segmentation result as the second level set, H (φ) Namely zero level collection.
Level set loss function mainly includes two parts, and first part is first level collection φpreShape, second It is divided into the shape of zero level collection H (φ), specifically, the calculation formula of first part is as follows:
Wherein, φgtIt is true value level set, that is, the true value of input picture is as level set.
The calculation formula of second part is as follows:
Wherein, H (φpre) indicate the probability y that U-net network exportsi, yi' indicate probability yiCorresponding true value.
Further, the calculation formula of the level set loss function of minimum is as follows:
Wherein, C is active contour (boundary of i.e. above-mentioned target object), and last two (i.e. above-mentioned for join domain w Target object region) inside and outside, c1And c2It is constant, can be calculated by following formula:
In addition, including two kinds of true value, i.e. true value level set and true value label in U-net network training process, respectively as schemed Shown in 4a and Fig. 4 b.True value level set can be converted to by quick distance, and due to using Euclidean distance, image becomes Greatly, whole distance becomes larger, if directly normalizing to [- 1,1], the response on corresponding boundary can be reduced, therefore can pass through As under type converts true value level set:
Wherein, φNgt(x, y) is the level set (i.e. above-mentioned treated second function value) after conversion.
That is, the functional value of true value level set can be compared with preset value, if it is larger than or equal to a preset value Functional value can be then converted to the preset value by (i.e. the first preset value);If it is less than another preset value, (i.e. first is default Value), then functional value can be converted to the preset value;For other situations, then remain unchanged.It may further will be after conversion Level set normalizes to [- 1,1], and the result after normalization is as illustrated in fig. 4 c.
Optionally, in step S104, image segmentation is carried out to the first image using U-shaped full convolutional neural networks model, is obtained To before the boundary of target object, this method further includes following steps: the first image is handled using edge detection algorithm, Obtain the second image of target object region;Image point is carried out to the second image using U-shaped full convolutional neural networks model It cuts, obtains the boundary of target object.
Specifically, in order to reduce extra interference information, edge detection algorithm can be used and extract crystalline lens administrative division map Picture, obtains the second image, and image size is 1024*1024, and by the image extracted be input to the full convolutional neural networks of U-shaped into Row image segmentation.
Optionally, the first image is handled using edge detection algorithm, obtains the second of target object region Image, comprising: the first image is handled using multistage edge detection algorithm, obtains the second image.
Specifically, the crystalline body region of multistage edge detection algorithm canny operator extraction can be used as pretreatment.
Optionally, the first image comprising target object is obtained, comprising: utilize leading portion means of optical coherence tomography pair Target object is scanned, and obtains the first image.
Specifically, crystalline lens can be shot by AS-OCT technology, obtains the eyeground AS-OCT figure (i.e. above-mentioned One image).
Through the above scheme, compared with prior art, the present invention directly models edge, to the edge to learn There is more preceding control force;It is easy addition of constraints condition, many constraint conditions such as shape or region consistency are easy to level Set representations have been obviously improved effect;Generalization is strong, is readily migrate into other tasks.
Embodiment 2
According to embodiments of the present invention, a kind of embodiment of the identification device of target object is provided.
Fig. 5 is a kind of schematic diagram of the identification device of target object according to an embodiment of the present invention, as shown in figure 5, the dress Set includes: to obtain module 52 and image segmentation module 54.
Wherein, module 52 is obtained for obtaining the first image comprising target object;Image segmentation module 54 is used to utilize U The full convolutional neural networks model of shape carries out image segmentation to the first image, obtains the boundary of target object, wherein the full convolution of U-shaped Neural network model uses level set loss function.
Optionally, above-mentioned target object can be lens nucleus.
Specifically, the first above-mentioned image can be the eyeground AS-OCT figure, include Phakic in image as shown in Figure 2 a Structure, wherein white box region is crystalline body region.Due to being obscured between lens nucleus and the boundary of cortex, Ke Yitong It crosses and the boundary of lens nucleus and cortex is accurately divided, realize the accurate segmentation to lens structure, hence, it can be determined that Target object is lens nucleus, and target object region is as shown in Figure 2 b, wherein the part of black is cortex, among black White portion be core.
In medical image segmentation field, quickly grown currently based on deep learning image segmentation algorithm, especially U-shaped is complete Convolutional neural networks model, the i.e. appearance of U-net network, in the field of medical imaging such as Lung neoplasm, tumour and blood vessel thickness segmentations Achieve biggish development.It therefore, can be by U-net network as the initial profile of lens structure in AS-OCT image Segmentation, determines lens nucleus region, final segmentation result is as shown in Figure 2 c.
It should be noted that carrying out image segmentation to the eyeground AS-OCT figure by U-net network, crystalline substance can be not only determined Shape body core region, can also determine lenticular cornea and cortical area.
Due to the irregularity boundary that U-net network is partitioned into, the especially segmentation of lens nucleus, and Level Set Method is in shape Shape, noise resolution ratio and block etc. and to obtain preferable performance under some constraints, therefore, can be damaged by e-learning level set Lose function, the final segmentation result of the zero level set representations finally obtained, that is, another horizontal set identifier target object region Boundary.
Above-described embodiment through the invention, after getting the first image comprising target object by acquisition module, It can be utilized using the full convolutional neural networks model of U-shaped of level set loss function by image segmentation module to the first image Image segmentation is carried out, obtains the boundary of target object, namely obtain final segmentation result.It is easily noted that, passes through knot It closes the full convolutional neural networks model of U-shaped and level set algorithm carries out lens structure segmentation, realize complete certainly based on deep learning Dynamic lens structure segmentation has reached the technical effect of the accuracy and repeatability that effectively improve lens structure segmentation, into And the recognition accuracy for solving target object in the prior art is lower, and the technical problem that repeatability is poor.
Optionally, the full convolutional neural networks model of U-shaped includes: encoding model and decoded model, and encoding model includes: multiple First network block, first network block include: sequentially connected at least two first convolutional layer, and the first convolutional layer is linear using amendment Unit activating function and pondization operate, in the last one first convolutional layer and next first network block in first network block First the first convolutional layer connection;Decoded model includes: multiple second network blocks and output layer, first network block and the second network The quantity of block is identical, and the second network block includes: sequentially connected cascading layers and at least two second convolutional layers, cascading layers with it is corresponding First network block connected with the last one second convolutional layer in upper second network block, cascading layers and corresponding first net Network block is connected using duplication with combineds jump connection type, the last one in output layer and the last one second network block the Two side output layer connection, output layer use level set loss function.
Optionally, above-mentioned pondization operation may include one of following: maximum pondization operation and the operation of mean value pondization.
Optionally, level set loss function passes through lower unit such as and determines: determining submodule, acquisition submodule and processing submodule Block.
Wherein it is determined that submodule is used to determine that the result of input to output layer to be first level collection, and determine that output layer is defeated Result out is the second level set;Acquisition submodule is for obtaining the shape of first level collection and the shape of the second level set;Place The state that submodule is used for shape and the second level set based on first level collection is managed, level set loss function is obtained.
Optionally, acquisition submodule includes: first acquisition unit and first processing units.
Wherein, first acquisition unit is used to obtain the first function value of first level collection and the true value water of the first image The second function value of flat collection;First processing units are used to be based on first function value and second function value, obtain first level collection Shape.
Specifically, first processing units include: the first acquisition subelement, the second acquisition subelement, third acquisition subelement Subelement is obtained with the 4th.
Wherein, the first acquisition subelement is used to obtain the difference of second function value and first function value, obtains difference;Second obtains It takes subelement for obtaining the absolute value of difference, obtains absolute value;Third obtains square that subelement is used to obtain absolute value, obtains To square value;4th acquisition subelement is used to obtain the integral of square value, obtains the shape of first level collection.
In addition, the device further include: the first processing subelement, second processing subelement and third handle subelement.
Wherein, the first processing subelement is used to handle second function value based on level set transfer function, obtains everywhere Second function value after reason;Second processing subelement is returned for treated, second function value to be normalized One changes functional value;Third handles subelement and is used to be based on first function value and normalized function value, obtains the shape of first level collection Shape.
Similarly, acquisition submodule includes: second acquisition unit, the second processing unit and third acquiring unit.
Wherein, second acquisition unit is used to obtain multiple probability values of input to output layer and each probability value corresponds to True value;The second processing unit is used to be based on each probability value and corresponding true value, obtains the product of each probability value;Third obtains It takes unit for obtaining the sum of products of multiple probability values, obtains the shape of the second level set.
Specifically, the second processing unit includes: that the 5th acquisition subelement and the 6th obtain subelement.
Wherein, the 5th acquisition subelement is used to obtain the logarithm of each probability value, obtains the logarithm of each probability value;The Six acquisition subelements obtain multiplying for each probability value for obtaining the logarithm of each probability value and the product of corresponding true value Product.
Optionally, processing submodule includes: the 4th acquiring unit, the 5th acquiring unit, the 6th acquiring unit, the 7th acquisition Unit and the 8th acquiring unit.
Wherein, the 4th acquiring unit obtains first and multiplies for obtaining the shape of first level collection and the product of the first parameter Product;5th acquiring unit obtains the second product for obtaining the shape of the second level set and the product of the second parameter;6th obtains Unit is used to obtain the length on the boundary of target object and the product of third parameter, obtains third product;7th acquiring unit is used In the product for the integrated value and the 4th parameter for obtaining target object region, the 4th product is obtained, wherein integrated value is used for table Levy the inside and outside of linking objective object region;8th acquiring unit is for obtaining the first product, the second product, third The 4th sum of products of sum of products, obtains level set loss function.
Optionally, which includes: processing module.
Wherein, processing module obtains target object place for handling using edge detection algorithm the first image Second image in region;Image segmentation module is also used to carry out image to the second image using the full convolutional neural networks model of U-shaped Segmentation, obtains the boundary of target object.
Optionally, processing module includes: processing submodule.
Wherein, processing submodule is also used to handle the first image using multistage edge detection algorithm, obtains second Image.
Optionally, obtaining module includes: scanning submodule.
Wherein, scanning submodule is obtained for being scanned using leading portion means of optical coherence tomography to target object To the first image.
Embodiment 3
According to embodiments of the present invention, a kind of embodiment of storage medium is provided, storage medium includes the program of storage, In, in program operation, equipment where control storage medium executes the recognition methods of the target object in above-described embodiment 1.
Embodiment 4
According to embodiments of the present invention, a kind of embodiment of processor is provided, processor is for running program, wherein journey The recognition methods of the target object in above-described embodiment 1 is executed when sort run.
The serial number of the above embodiments of the invention is only for description, does not represent the advantages or disadvantages of the embodiments.
In the above embodiment of the invention, it all emphasizes particularly on different fields to the description of each embodiment, does not have in some embodiment The part of detailed description, reference can be made to the related descriptions of other embodiments.
In several embodiments provided herein, it should be understood that disclosed technology contents can pass through others Mode is realized.Wherein, the apparatus embodiments described above are merely exemplary, such as the division of the unit, Ke Yiwei A kind of logical function partition, there may be another division manner in actual implementation, for example, multiple units or components can combine or Person is desirably integrated into another system, or some features can be ignored or not executed.Another point, shown or discussed is mutual Between coupling, direct-coupling or communication connection can be through some interfaces, the INDIRECT COUPLING or communication link of unit or module It connects, can be electrical or other forms.
The unit as illustrated by the separation member may or may not be physically separated, aobvious as unit The component shown may or may not be physical unit, it can and it is in one place, or may be distributed over multiple On unit.It can some or all of the units may be selected to achieve the purpose of the solution of this embodiment according to the actual needs.
It, can also be in addition, the functional units in various embodiments of the present invention may be integrated into one processing unit It is that each unit physically exists alone, can also be integrated in one unit with two or more units.Above-mentioned integrated list Member both can take the form of hardware realization, can also realize in the form of software functional units.
If the integrated unit is realized in the form of SFU software functional unit and sells or use as independent product When, it can store in a computer readable storage medium.Based on this understanding, technical solution of the present invention is substantially The all or part of the part that contributes to existing technology or the technical solution can be in the form of software products in other words It embodies, which is stored in a storage medium, including some instructions are used so that a computer Equipment (can for personal computer, server or network equipment etc.) execute each embodiment the method for the present invention whole or Part steps.And storage medium above-mentioned includes: that USB flash disk, read-only memory (ROM, Read-Only Memory), arbitrary access are deposited Reservoir (RAM, Random Access Memory), mobile hard disk, magnetic or disk etc. be various to can store program code Medium.
The above is only a preferred embodiment of the present invention, it is noted that for the ordinary skill people of the art For member, various improvements and modifications may be made without departing from the principle of the present invention, these improvements and modifications are also answered It is considered as protection scope of the present invention.

Claims (17)

1. a kind of recognition methods of target object characterized by comprising
Obtain the first image comprising target object;
Image segmentation is carried out to the first image using U-shaped full convolutional neural networks model, obtains the side of the target object Boundary, wherein the full convolutional neural networks model of U-shaped uses level set loss function.
2. the method according to claim 1, wherein the full convolutional neural networks model of the U-shaped includes: coding mould Type and decoded model,
The encoding model includes: multiple first network blocks, and the first network block includes: sequentially connected at least two first Convolutional layer, first convolutional layer are operated using the linear unit activating function of amendment and pondization, in the first network block most The first convolutional layer of the latter is connect with first the first convolutional layer in next first network block;
The decoded model includes: multiple second network blocks and output layer, the first network block and second network block Quantity is identical, and second network block includes: sequentially connected cascading layers and at least two second convolutional layers, the cascading layers with Corresponding first network block is connected with the last one second convolutional layer in upper second network block, the cascading layers with it is described Corresponding first network block is connected using duplication with combined jump connection type, the output layer and the last one second network The last one second side output layer connection in block, the output layer use the level set loss function.
3. according to the method described in claim 2, it is characterized in that, pondization operation includes one of following: maximum Chi Huacao Make and mean value pondization operates.
4. according to the method described in claim 2, it is characterized in that, the level set loss function is determined as follows:
It determines that the result for being input to the output layer is first level collection, and determines that the result of the output layer output is the second water Flat collection;
Obtain the shape of the first level collection and the shape of second level set;
The state of shape and second level set based on the first level collection, obtains the level set loss function.
5. according to the method described in claim 4, it is characterized in that, obtaining the shape of the first level collection, comprising:
Obtain the second function value of the first function value of the first level collection and the true value level set of the first image;
Based on the first function value and the second function value, the shape of the first level collection is obtained.
6. according to the method described in claim 5, it is characterized in that, be based on the first function value and the second function value, Obtain the shape of the first level collection, comprising:
The difference for obtaining the second function value and the first function value, obtains difference;
The absolute value for obtaining the difference, obtains absolute value;
Square for obtaining the absolute value, obtains square value;
The integral for obtaining the square value obtains the shape of the first level collection.
7. according to the method described in claim 5, it is characterized in that, obtain true value level set second function value after, institute State method further include:
The second function value is handled based on level set transfer function, the second function value that obtains that treated;
Treated that second function value is normalized to described, obtains normalized function value;
Based on the first function value and the normalized function value, the shape of the first level collection is obtained.
8. according to the method described in claim 4, it is characterized in that, obtaining the shape of second level set, comprising:
Obtain the multiple probability values and the corresponding true value of each probability value for being input to the output layer;
Based on each probability value and corresponding true value, the product of each probability value is obtained;
The sum of products for obtaining the multiple probability value obtains the shape of second level set.
9. according to the method described in claim 8, it is characterized in that, being obtained described based on each probability value and corresponding true value The product of each probability value, comprising:
The logarithm for obtaining each probability value obtains the logarithm of each probability value;
The logarithm of each probability value and the product of corresponding true value are obtained, the product of each probability value is obtained.
10. according to the method described in claim 4, it is characterized in that, the shape based on the first level collection and described second The state of level set obtains the level set loss function, comprising:
The shape of the first level collection and the product of the first parameter are obtained, the first product is obtained;
The shape of second level set and the product of the second parameter are obtained, the second product is obtained;
The length on the boundary of the target object and the product of third parameter are obtained, third product is obtained;
The integrated value of the target object region and the product of the 4th parameter are obtained, the 4th product is obtained, wherein the product Score value is used to characterize the inside and outside of the connection target object region;
First product, second product, the 4th sum of products described in the third sum of products are obtained, the level is obtained Collect loss function.
11. the method according to claim 1, wherein using the full convolutional neural networks model of U-shaped to described the One image carries out image segmentation, before obtaining the boundary of the target object, the method also includes:
The first image is handled using edge detection algorithm, obtains the second figure of the target object region Picture;
Image segmentation is carried out to second image using U-shaped full convolutional neural networks model, obtains the side of the target object Boundary.
12. according to the method for claim 11, which is characterized in that carried out using edge detection algorithm to the first image Processing, obtains the second image of the target object region, comprising:
The first image is handled using multistage edge detection algorithm, obtains second image.
13. the method according to claim 1, wherein obtaining the first image comprising target object, comprising:
The target object is scanned using leading portion means of optical coherence tomography, obtains the first image.
14. the method according to claim 1, wherein the target object is lens nucleus.
15. a kind of identification device of target object characterized by comprising
Module is obtained, for obtaining the first image comprising target object;
Image segmentation module is obtained for carrying out image segmentation to the first image using the full convolutional neural networks model of U-shaped The boundary of the target object, wherein the full convolutional neural networks model of U-shaped uses level set loss function.
16. a kind of storage medium, which is characterized in that the storage medium includes the program of storage, wherein run in described program When control the storage medium where equipment perform claim require any one of 1 to 14 described in target object identification side Method.
17. a kind of processor, which is characterized in that the processor is for running program, wherein right of execution when described program is run Benefit require any one of 1 to 14 described in target object recognition methods.
CN201910176768.7A 2019-03-08 2019-03-08 Target object identification method and device Active CN109919098B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910176768.7A CN109919098B (en) 2019-03-08 2019-03-08 Target object identification method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910176768.7A CN109919098B (en) 2019-03-08 2019-03-08 Target object identification method and device

Publications (2)

Publication Number Publication Date
CN109919098A true CN109919098A (en) 2019-06-21
CN109919098B CN109919098B (en) 2021-06-15

Family

ID=66963937

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910176768.7A Active CN109919098B (en) 2019-03-08 2019-03-08 Target object identification method and device

Country Status (1)

Country Link
CN (1) CN109919098B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111539306A (en) * 2020-04-21 2020-08-14 中南大学 Remote sensing image building identification method based on activation expression replaceability
CN117422880A (en) * 2023-12-18 2024-01-19 齐鲁工业大学(山东省科学院) Segmentation method and system combining improved attention mechanism and CV model

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102831608A (en) * 2012-08-06 2012-12-19 哈尔滨工业大学 Unsteady measurement algorithm based image segmentation method of improved rule distance level set
US20170109883A1 (en) * 2015-10-19 2017-04-20 The Charles Stark Draper Laboratory, Inc. System and method for the segmentation of optical coherence tomography slices
CN108492286A (en) * 2018-03-13 2018-09-04 成都大学 A kind of medical image cutting method based on the U-shaped convolutional neural networks of binary channel
CN109035252A (en) * 2018-06-29 2018-12-18 山东财经大学 A kind of super-pixel method towards medical image segmentation
CN109087327A (en) * 2018-07-13 2018-12-25 天津大学 A kind of thyroid nodule ultrasonic image division method cascading full convolutional neural networks

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102831608A (en) * 2012-08-06 2012-12-19 哈尔滨工业大学 Unsteady measurement algorithm based image segmentation method of improved rule distance level set
US20170109883A1 (en) * 2015-10-19 2017-04-20 The Charles Stark Draper Laboratory, Inc. System and method for the segmentation of optical coherence tomography slices
CN108492286A (en) * 2018-03-13 2018-09-04 成都大学 A kind of medical image cutting method based on the U-shaped convolutional neural networks of binary channel
CN109035252A (en) * 2018-06-29 2018-12-18 山东财经大学 A kind of super-pixel method towards medical image segmentation
CN109087327A (en) * 2018-07-13 2018-12-25 天津大学 A kind of thyroid nodule ultrasonic image division method cascading full convolutional neural networks

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
PENGSHUAI YIN ET AL: ""Automatic Segmentation o f Cortex and Nucleus in Anterior Segment OCT Images"", 《COMPUTATIONAL PATHOLOGY AND OPHTHALMIC MEDICAL IMAGE ANALYSIS》 *
VENI G, ET AL: ""Echocardiography segmentation based on a shape-guided deformable model driven by a fully convolutional network prior"", 《2018 IEEE 15TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI 2018)》 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111539306A (en) * 2020-04-21 2020-08-14 中南大学 Remote sensing image building identification method based on activation expression replaceability
CN111539306B (en) * 2020-04-21 2021-07-06 中南大学 Remote sensing image building identification method based on activation expression replaceability
CN117422880A (en) * 2023-12-18 2024-01-19 齐鲁工业大学(山东省科学院) Segmentation method and system combining improved attention mechanism and CV model
CN117422880B (en) * 2023-12-18 2024-03-22 齐鲁工业大学(山东省科学院) Segmentation method and system combining improved attention mechanism and CV model

Also Published As

Publication number Publication date
CN109919098B (en) 2021-06-15

Similar Documents

Publication Publication Date Title
CN111340819B (en) Image segmentation method, device and storage medium
US20230360313A1 (en) Autonomous level identification of anatomical bony structures on 3d medical imagery
JP6842481B2 (en) 3D quantitative analysis of the retinal layer using deep learning
CN108416776A (en) Image-recognizing method, pattern recognition device, computer product and readable storage medium storing program for executing
SG190730A1 (en) Method and an apparatus for determining vein patterns from a colour image
CN106815853A (en) To the dividing method and device of retinal vessel in eye fundus image
CN111862009B (en) Classifying method of fundus OCT (optical coherence tomography) images and computer readable storage medium
CN108764342B (en) Semantic segmentation method for optic discs and optic cups in fundus image
CN108932536A (en) Human face posture method for reconstructing based on deep neural network
CN107423721A (en) Interactive action detection method, device, storage medium and processor
CN109919954A (en) The recognition methods of target object and device
CN109919932A (en) The recognition methods of target object and device
CN110070531A (en) For detecting the model training method of eyeground picture, the detection method and device of eyeground picture
CN111462049A (en) Automatic lesion area form labeling method in mammary gland ultrasonic radiography video
CN102567734A (en) Specific value based retina thin blood vessel segmentation method
CN109934813A (en) The recognition methods of target object and device
CN110570394A (en) medical image segmentation method, device, equipment and storage medium
CN109919098A (en) The recognition methods of target object and device
CN115393239A (en) Multi-mode fundus image registration and fusion method and system
CN113920109A (en) Medical image recognition model training method, recognition method, device and equipment
WO2015143435A1 (en) Graph search using non-euclidean deformed graph
CN105976321B (en) A kind of optical coherence tomography image super-resolution rebuilding method and device
CN106504199B (en) A kind of eye fundus image Enhancement Method and system
CN114782337A (en) OCT image recommendation method, device, equipment and medium based on artificial intelligence
Kim et al. Facial landmark extraction scheme based on semantic segmentation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant