WO2021179205A1 - 医学图像分割方法、医学图像分割装置及终端设备 - Google Patents
医学图像分割方法、医学图像分割装置及终端设备 Download PDFInfo
- Publication number
- WO2021179205A1 WO2021179205A1 PCT/CN2020/078800 CN2020078800W WO2021179205A1 WO 2021179205 A1 WO2021179205 A1 WO 2021179205A1 CN 2020078800 W CN2020078800 W CN 2020078800W WO 2021179205 A1 WO2021179205 A1 WO 2021179205A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- medical image
- segmentation
- output
- result
- model
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 65
- 238000003709 image segmentation Methods 0.000 title claims abstract description 47
- 230000011218 segmentation Effects 0.000 claims abstract description 230
- 238000012545 processing Methods 0.000 claims abstract description 59
- 230000004927 fusion Effects 0.000 claims abstract description 26
- 239000011159 matrix material Substances 0.000 claims description 136
- 238000012549 training Methods 0.000 claims description 31
- 230000006870 function Effects 0.000 claims description 27
- 238000004590 computer program Methods 0.000 claims description 24
- 230000008569 process Effects 0.000 claims description 18
- 238000005070 sampling Methods 0.000 claims description 14
- 238000004364 calculation method Methods 0.000 claims description 12
- 238000013527 convolutional neural network Methods 0.000 claims description 12
- 230000004913 activation Effects 0.000 claims description 9
- 238000013528 artificial neural network Methods 0.000 claims description 3
- 238000010586 diagram Methods 0.000 description 8
- 238000005314 correlation function Methods 0.000 description 5
- 238000005516 engineering process Methods 0.000 description 4
- 238000012360 testing method Methods 0.000 description 4
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 230000000694 effects Effects 0.000 description 3
- 230000003190 augmentative effect Effects 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 238000013136 deep learning model Methods 0.000 description 2
- 238000002372 labelling Methods 0.000 description 2
- 238000012546 transfer Methods 0.000 description 2
- 230000009466 transformation Effects 0.000 description 2
- 238000012795 verification Methods 0.000 description 2
- 238000005481 NMR spectroscopy Methods 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000002583 angiography Methods 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 238000004422 calculation algorithm Methods 0.000 description 1
- 238000002591 computed tomography Methods 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000002059 diagnostic imaging Methods 0.000 description 1
- 235000019800 disodium phosphate Nutrition 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000003902 lesion Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 210000000056 organ Anatomy 0.000 description 1
- 238000007500 overflow downdraw method Methods 0.000 description 1
- 238000002600 positron emission tomography Methods 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 238000002604 ultrasonography Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
Definitions
- This application relates to the field of image segmentation technology, and in particular to medical image segmentation methods, medical image segmentation devices, terminal equipment, and computer-readable storage media.
- Medical image segmentation is a key step in medical image processing and analysis.
- information technology and high-end medical imaging technology represented by artificial intelligence have continued to develop, and the application of deep learning in the field of medical image segmentation has also received more and more attention.
- the embodiments of the present application provide a medical image segmentation method, a medical image segmentation device, a terminal device, and a computer-readable storage medium, which can improve the accuracy of image segmentation of a medical image.
- the first aspect of the embodiments of the present application provides a medical image segmentation method, including:
- the medical image to be detected is input into a trained segmentation model, wherein the segmentation model includes an encoder and a decoder, the encoder includes a plurality of first hierarchical structures, and the plurality of first hierarchical structures includes a first An input layer, at least one first intermediate layer, and a first output layer, the decoder includes a second input layer, at least one second intermediate layer, and a second output layer, wherein the input of any second intermediate layer includes the first 2.
- the medical image to be detected is processed by the segmentation model to obtain an output result of the segmentation model, wherein the output result includes a segmentation result of the medical feature region in the medical image to be detected.
- a second aspect of the embodiments of the present application provides a medical image segmentation device.
- the medical image segmentation device may include a module for implementing the steps of the medical image segmentation method described above.
- a third aspect of the embodiments of the present application provides a computer-readable storage medium, the computer-readable storage medium stores computer-readable instructions, and when the computer-readable instructions are executed by a processor, the processor executes the foregoing Steps of medical image segmentation method.
- a fourth aspect of the embodiments of the present application provides a computer device, which includes a memory and a processor.
- the memory stores computer-readable instructions.
- the processor implements the above-mentioned medical image segmentation when the computer-readable instructions are executed. Method steps.
- FIG. 1 is a schematic flowchart of a medical image segmentation method provided by an embodiment of the present application
- Fig. 2 is an exemplary structure of the segmentation model provided by an embodiment of the present application
- FIG. 3 is a schematic flowchart of step S103 according to an embodiment of the present application.
- FIG. 4 is an exemplary schematic diagram of performing second processing on the first feature matrix by the weight obtaining module according to an embodiment of the present application
- FIG. 5 is an exemplary schematic diagram of the segmentation model and the discrimination model provided by an embodiment of the present application.
- Fig. 6 is a schematic structural diagram of a medical image segmentation device provided by an embodiment of the present application.
- FIG. 7 is a schematic structural diagram of a terminal device provided by an embodiment of the present application.
- the medical image segmentation method provided by the embodiments of this application can be applied to servers, desktop computers, mobile phones, tablet computers, wearable devices, vehicle-mounted devices, augmented reality (AR)/virtual reality (VR) devices, and notebooks.
- terminal devices such as computers, ultra-mobile personal computers (UMPC), netbooks, and personal digital assistants (PDAs)
- UMPC ultra-mobile personal computers
- PDAs personal digital assistants
- FIG. 1 shows a flowchart of a medical image segmentation method provided by an embodiment of the present application, and the medical image segmentation method may be applied to terminal equipment.
- the medical image segmentation method may include:
- Step S101 Obtain a medical image to be detected.
- the type and acquisition method of the medical image to be detected are not limited here.
- the medical image to be detected may include one or more of endoscopic images, angiography images, computed tomography images, positron emission tomography images, nuclear magnetic resonance images, and ultrasound images.
- the medical image to be detected often includes a medical characteristic area, where the medical characteristic area may be, for example, a lesion area, a specific tissue or organ area, and so on.
- Step S102 Input the medical image to be detected into a trained segmentation model, where the segmentation model includes an encoder and a decoder, the encoder includes a plurality of first hierarchical structures, and the plurality of first hierarchical structures It includes a first input layer, at least one first intermediate layer, and a first output layer.
- the decoder includes a second input layer, at least one second intermediate layer, and a second output layer.
- the input of any second intermediate layer includes The fusion result of the output of the previous layer of the second intermediate layer and the output of the at least two first hierarchical structures.
- the trained segmentation model may be used to perform image segmentation on the medical image to be detected, so as to obtain information such as the contour of the medical feature region in the medical image to be detected.
- the trained segmentation model may include an encoder and a decoder, wherein the specific structure of the encoder and the decoder may be determined based on an existing or future machine learning model.
- the structure of the encoder and the decoder may be symmetrical.
- the number of first-level structures in the encoder is the same as the number of second-level structures included in the decoder.
- the number of the first hierarchical structure can be determined according to actual requirements.
- the first hierarchical structure may have 5 layers.
- the encoder may include 3 first intermediate layers.
- any one of the first hierarchical structures may include one or more sub-layers.
- any hierarchical structure in the encoder may include a convolutional layer and a down-sampling layer.
- decoding The second hierarchical structure corresponding to the first hierarchical structure in the device may include an up-sampling layer and a convolutional layer.
- the output of any first hierarchical structure in the encoder may be the output of the down-sampling layer in the first hierarchical structure.
- the trained segmentation model may be improved based on an existing model such as U-Net.
- the existing U-Net model is designed based on a jump-connected full convolutional network, which includes an encoder and decoder with a symmetric structure.
- the encoder and decoder of the existing U-Net model There is a one-to-one corresponding intermediate layer.
- the output of the intermediate layer of the encoder can be transferred to the corresponding intermediate layer in the decoder, and after the transfer, it is spliced and fused with the output of the previous layer of the corresponding intermediate layer in the decoder , And use the result of the splicing and fusion as the input of the corresponding middle layer in the decoder.
- the input of any second intermediate layer in the decoder may include the fusion result of the output of the previous layer of the second intermediate layer and the output of at least two first-level structures.
- the input of any second intermediate layer in the decoder may include the output of the previous layer of the second intermediate layer, the output of the first hierarchical structure corresponding to the second intermediate layer, and the output of the first layer corresponding to the second intermediate layer.
- the decoder can obtain the extracted features of different scales of multiple first-level structures, thereby fusing the multi-scale features to make full use of the context information of the pixels in the medical image to make the medical image segmentation.
- the output of the previous layer of the second intermediate layer may be spliced with the output of at least two first-level structures to obtain the fusion result.
- the fusion result of the output of the previous layer of the second intermediate layer and the output of the at least two first hierarchical structures is the output of the previous layer of the second intermediate layer and the output of the second intermediate layer.
- the second intermediate layer can obtain and make full use of feature information of different depths through multiple jump connections from the encoder to the decoder, thereby improving the efficiency of feature expression and improving the segmentation model The segmentation performance.
- the encoder of the segmentation network may have a first-level structure of 5 layers, namely A, B, C, D, and E.
- the structure of the decoder is symmetrical to the encoder, that is, the A, B, C, D, and E layers of the encoder correspond to the A', B', C', D', and E'layers in the decoder, then
- the A'layer in the decoder can obtain the fusion result of the output of the A layer of the encoder and the output of the B'layer of the decoder, and the B'layer in the decoder can obtain the output of the A layer and the B layer of the encoder.
- the output is fused with the output of the decoder C'layer.
- the C'layer in the decoder can obtain the output of the B layer of the encoder, the output of the C layer and the output of the decoder D'layer, and so on .
- FIG. 2 is only an exemplary structure of the segmentation model, and is not a limitation.
- step S103 the medical image to be detected is processed by the segmentation model to obtain an output result of the segmentation model, where the output result includes a segmentation result of the medical feature region in the medical image to be detected.
- the segmentation model may output a segmentation result about the medical feature region in the medical image to be detected, wherein, specifically, the segmentation result includes contour information of the medical feature region.
- the segmentation model includes a weight acquisition module, and the weight acquisition module is located between the encoder and the decoder;
- the processing the medical image to be detected by the segmentation model to obtain the output result of the segmentation model includes:
- Step S301 Perform first processing on the medical image to be detected by the encoder to obtain a first feature matrix output by the encoder;
- Step S302 input the first feature matrix to the weight acquisition module
- Step S303 Perform a second process on the first feature matrix by the weight acquisition module to obtain the weight matrix output by the weight acquisition module;
- Step S304 fusing the weight matrix and the first feature matrix to obtain a second feature matrix
- Step S305 Based on the second feature matrix, perform third processing by the decoder to obtain the output result.
- the weight acquisition module may be set between the encoder and the decoder, thereby using the attention mechanism to improve the segmentation model's ability to represent the segmented region .
- the weight obtaining module may perform a second processing on the first feature matrix through a preset correlation function, so as to obtain the weight matrix output by the weight obtaining module.
- the correlation function may be obtained by combining convolution operation and activation function, or may be obtained by combining multiplication, addition, and other specific functions.
- each element in the weight matrix may respectively represent the weight value of the corresponding element in the corresponding first feature matrix.
- the weight matrix and the first feature matrix can be fused in multiple ways. For example, the weight matrix can be added to the first feature matrix, or the corresponding elements in the weight matrix can be combined with the first feature matrix. The corresponding elements in the first feature matrix are multiplied; in addition, the fusion may also include matrix dimension transformation, etc., for example, the weight matrix, the first feature matrix, or both may be added after The obtained matrix is dimensionally transformed to obtain the second feature matrix.
- the performing the second processing on the first feature matrix by the weight obtaining module to obtain the weight matrix output by the weight obtaining module includes:
- performing the first convolution processing on the first feature matrix may be performing a convolution operation on the first convolution matrix and the first feature matrix, wherein the first convolution matrix may have a dimension of 1*1 matrix;
- the performing the second convolution processing on the first feature matrix may be performing a convolution operation on the second convolution matrix and the first feature matrix, wherein the second convolution matrix It can be a matrix with a dimension of 1*1.
- the activation function may be a Softmax activation function or the like.
- the above-mentioned second processing performed on the first feature matrix by the weight obtaining module may be represented by an association function.
- the element at each position (x i , x j ) in the first feature matrix can be calculated through the correlation function f(x i , x j ) to obtain the corresponding
- the weight value, the correlation function f(x i , x j ) can be expressed as:
- ⁇ (x i ) is the first convolution process performed by the first embedding layer in the weight obtaining module
- ⁇ (x j ) is the second convolution process performed by the second embedding layer in the weight obtaining module.
- the third processing performed by the decoder based on the second feature matrix to obtain the output result includes:
- the decoder Based on the sixth feature matrix, the decoder performs third processing to obtain the output result.
- FIG. 4 it is an exemplary schematic diagram of performing the second processing on the first feature matrix by the weight obtaining module.
- the first convolution processing may be the performing the first convolution processing on the first feature matrix, which may be convolving a first convolution matrix with a dimension of 1*1 with the first feature matrix Operation; said performing the second convolution processing on the first feature matrix may be performing a convolution operation on the second convolution matrix with a dimension of 1*1 and the first feature matrix.
- the activation function may be a Softmax activation function.
- the second feature matrix may be further combined with the first output
- the output of the previous layer of the layer is fused to obtain the sixth feature matrix, and then the sixth feature matrix is input to the decoder.
- the second input layer of the decoder can obtain the encoder
- the feature information extracted from different depths can be fused, so that the second input layer of the encoder can better utilize some context information in the medical image for processing, thereby improving The segmentation performance of the segmentation model.
- the method before inputting the medical image to be detected into the trained segmentation model, the method further includes:
- the segmentation model to be trained is trained through the discriminant model until the training is completed, and the trained segmentation model is obtained, wherein the input of the discriminant model includes at least part of the output of the segmentation model to be trained, and the discriminant model includes volume A product neural network and an up-sampling layer, the output of the convolutional neural network is the input of the up-sampling layer.
- training is based on the form of generating a confrontation network, so that a small amount of labeled medical image data and a large amount of unlabeled medical image data can be used for training, thereby reducing The dependence on the large number of finely labeled medical image data is reduced, and the training performance is improved.
- the discriminant model may include a convolutional neural network and an up-sampling layer, where the up-sampling layer may be used to output a confidence map, and the confidence map may be used to indicate the difference between each predicted segmentation result.
- the discriminant model may include other structures besides the convolutional neural network model and the upsampling layer, for example, the input or output of the convolutional neural network, or The input or output of the up-sampling layer is subjected to other processing, such as image enhancement processing, image binarization processing, and so on.
- the input of the discriminant model may include at least a part of the output of the segmentation model to be trained.
- it may also include real segmentation labels annotated with medical image samples to train the discriminant performance of the discriminant model.
- training the segmentation model to be trained by the discrimination model until the training is completed and obtaining the trained segmentation model includes:
- the medical image samples include labeled medical image samples and unlabeled medical image samples
- the labeled medical image samples are medical image samples labeled with real segmentation labels
- the unlabeled medical image samples are unlabeled medical image samples.
- the medical image sample may be obtained by normalizing the corresponding original medical image.
- the medical image sample may also be a medical image that has not been normalized.
- the confidence map may be used to indicate the location regions where the similarity of the real medical feature region corresponding to the real segmentation label in each of the predicted segmentation results meets the preset similarity condition.
- the loss value may include a cross-entropy loss related to the segmentation network, a supervision loss related to annotated medical image samples, a semi-supervised loss related to an unlabeled medical image sample, and/or a discrimination loss of the discriminant model, etc. One or more of etc.
- the semi-supervised loss regarding the unlabeled medical image sample may be determined based on the confidence map. Specifically, after the confidence map is obtained, the semi-supervised loss regarding the unlabeled medical image sample can be determined according to the confidence map. Among them, in some examples, the confidence map may be further processed, for example, the confidence map may be binarized or other encoding processing to highlight the credible region in the confidence map, so as to calculate the Semi-supervised loss of unlabeled medical image samples.
- the discriminant model and the to-be-trained segmentation model can be adjusted by means of backpropagation to update the gradient. Divide the parameters of the model until the loss value obtained meets the preset loss condition.
- Exemplary preset loss conditions may be that the loss value is less than the preset loss threshold, and conditions such as convergence.
- the correlation between the labeled medical image sample and the unlabeled medical image sample can be effectively used to obtain the relationship between the labeled medical image sample and the unlabeled medical image sample.
- the calculating the loss value of the segmentation model to be trained and the discrimination model based on a preset loss function according to the discrimination result, the confidence map, and the predicted segmentation result includes:
- the first predicted segmentation sub-result corresponding to the annotated medical image sample and the real segmentation label of the annotated medical image sample calculating a first loss value for the segmentation model
- the loss value is calculated according to the first loss value, the second loss value, and the third loss value.
- the first loss value may be calculated according to the first predicted segmentation sub-result corresponding to the labeled medical image sample and the real segmentation label of the labeled medical image sample, and the first loss value
- the specific calculation method and the type of loss function included can be determined based on actual experience.
- the first loss value may include a cross-entropy loss related to the segmentation network and/or a supervision loss related to annotated medical image samples, and so on.
- the third loss value may refer to the discrimination loss of the discrimination model.
- the second loss value may be calculated according to the second predicted segmentation sub-result corresponding to the unlabeled medical image sample and the confidence map.
- the second loss value may also be referred to as a semi-supervised loss.
- the first loss value, the second loss value, and the third loss value there may be multiple specific ways to calculate the loss value.
- the first loss value, the first loss value and the third loss value may be calculated.
- the weight values corresponding to the first loss value, the second loss value, and the third loss value may be preset, and the weight values corresponding to the first loss value and the second loss value may be set in advance.
- the weight value corresponding to the third loss value and the third loss value and the first loss value, the second loss value, and the third loss value are calculated to obtain the loss value.
- the calculating the second loss value of the segmentation model according to the second predicted segmentation sub-result corresponding to the unlabeled medical image sample in the predicted segmentation result and the confidence map includes:
- the encoding operation may be used to encode each position in the confidence map, and there may be multiple specific encoding methods, and the encoding may be used to encode each position in the confidence map.
- Carry out category labeling Exemplarily, when the category of each position in the confidence map includes two categories, the encoding operation may be a binarization operation; of course, the encoding operation may also include other encoding methods, for example, the encoding The calculation can be based on one-hot encoding (One-Hot Encoding) and other methods for encoding and so on.
- the encoded image may be used to annotate the credible region in the second predicted segmentation sub-result, wherein the credible region can be based on the authentic medical features in the second predicted segmentation sub-result.
- the similarity of the region is determined by the location region that meets the preset similarity condition.
- the relevant loss corresponding to the unlabeled medical image sample may be determined according to the credible region in the encoded image, that is, the second loss value.
- the following uses a specific example to illustrate an exemplary specific calculation method of the loss value in the embodiment of the present application.
- the labeled medical image sample is ⁇ I f , L f ⁇ , where L f is a true segmentation label of the labeled medical image sample.
- the unlabeled medical image sample is ⁇ I 0 ⁇ .
- the labeled medical image sample and the unlabeled medical image sample are input into the segmentation model to be trained, and the predicted segmentation result of each medical image sample by the segmentation model to be trained is obtained, wherein the first segment corresponding to the labeled medical image sample is obtained.
- the result of a predicted segmentation is
- the second predicted segmentation sub-result corresponding to the unlabeled medical image sample is S(L f ).
- the confidence map of the second predicted segmentation sub-result can be obtained through the up-sampling layer in the discriminant model At this time, according to the second predicted segmentation sub-result corresponding to the unlabeled medical image sample in the predicted segmentation result and the confidence map, the second loss value of the segmentation model is calculated
- I( ⁇ ) is an indicator function
- T semi is a preset confidence threshold
- Y( ⁇ I f ,L 0 ⁇ ) is used to indicate the coding category of the coded image corresponding to the confidence map
- the preset signal threshold T semi can be set according to actual experience or test results. By setting the preset signal threshold T semi , the sensitivity of model training can be controlled.
- the loss value may be calculated according to the first loss value, the second loss value, and the third loss value
- the ⁇ adv may be the supervision loss
- the ⁇ semi may be the second loss value The corresponding weight coefficient.
- the training results of the segmentation model and the discriminant model can be weighed and adjusted, for example, excessive correction can be avoided, and the effect such as cross entropy loss can be avoided.
- the cross entropy loss There may also be a weight coefficient corresponding to the third loss value.
- the discriminant model and the segmentation model to be trained are trained based on the loss value, until the obtained loss value meets the preset loss condition, it can also pass the medical image test for testing Samples and medical image verification samples for verification test and verify the segmentation model, so as to select the optimal segmentation model as the trained segmentation from the segmentation models whose loss values meet the preset loss conditions. Model.
- FIG. 5 it is an exemplary schematic diagram of the segmentation model and the discrimination model.
- the segmentation model and the discriminant model can use labeled medical image samples and unlabeled medical image samples to implement semi-supervised training.
- the medical image to be detected can be obtained, and the medical image to be detected can be input into a trained segmentation model, where the segmentation model includes an encoder and a decoder, and the encoder includes a plurality of first Hierarchical structure, the plurality of first hierarchical structures include a first input layer, at least one first intermediate layer, and a first output layer, and the decoder includes a second input layer, at least one second intermediate layer, and a second output layer , Wherein the input of any second intermediate layer includes the fusion result of the output of the previous layer of the second intermediate layer and the output of at least two first-level structures; at this time, it can pass through at least each of the decoders
- the middle layer acquires features of different scales extracted by multiple layers in the decoder, so as to make full use of the context information of the pixels in the medical image to perform medical image segmentation; to obtain the medical features in the medical image to be detected The segmentation result of the region.
- FIG. 6 shows a structural block diagram of a medical image segmentation device provided by an embodiment of the present application. part.
- the medical image segmentation device 6 includes:
- the first obtaining module 601 is used to obtain medical images to be detected
- the input module 602 is configured to input the medical image to be detected into a trained segmentation model, where the segmentation model includes an encoder and a decoder, the encoder includes a plurality of first-level structures, and the plurality of first-level structures
- the hierarchical structure includes a first input layer, at least one first intermediate layer, and a first output layer.
- the decoder includes a second input layer, at least one second intermediate layer, and a second output layer, wherein any second intermediate layer
- the input of includes the fusion result of the output of the previous layer of the second intermediate layer and the output of at least two first-level structures;
- the processing module 603 is configured to process the medical image to be detected by the segmentation model to obtain the output result of the segmentation model, wherein the output result includes information about the medical feature region in the medical image to be detected Segmentation result.
- the segmentation model includes a weight acquisition module, and the weight acquisition module is located between the encoder and the decoder;
- the processing module 603 specifically includes:
- a first processing unit configured to perform first processing on the medical image to be detected by the encoder to obtain a first feature matrix output by the encoder
- a first input unit configured to input the first feature matrix into the weight obtaining module
- a second processing unit configured to perform second processing on the first feature matrix by the weight acquisition module to obtain the weight matrix output by the weight acquisition module;
- a first fusion unit configured to fuse the weight matrix and the first feature matrix to obtain a second feature matrix
- the third processing unit is configured to perform third processing by the decoder based on the second feature matrix to obtain the output result.
- the second processing unit specifically includes:
- the first processing subunit is configured to perform first convolution processing on the first feature matrix to obtain a third feature matrix
- a second processing subunit configured to perform a second convolution process on the first feature matrix to obtain a fourth feature matrix
- the third processing subunit is configured to multiply the third feature matrix and the fourth feature matrix to obtain a fifth feature matrix
- the fourth processing subunit is configured to activate the fifth feature matrix through an activation function to obtain the weight matrix.
- the third processing unit specifically includes:
- the first fusion subunit is used to fuse the second feature matrix with the output of the previous layer of the first output layer to obtain a sixth feature matrix
- the first input subunit is used to input the sixth feature matrix to the decoder
- the fifth processing subunit is configured to perform third processing by the decoder based on the sixth feature matrix to obtain the output result.
- the medical image segmentation device 6 further includes:
- the training module is used to train the segmentation model to be trained through the discriminant model until the training is completed, and obtain the trained segmentation model, wherein the input of the discriminant model includes at least part of the output of the segmentation model to be trained, so
- the discriminant model includes a convolutional neural network and an upsampling layer, and the output of the convolutional neural network is the input of the upsampling layer.
- the training module specifically includes:
- An acquiring unit for acquiring medical image samples wherein the medical image samples include labeled medical image samples and unlabeled medical image samples, the labeled medical image samples are medical image samples labeled with real segmentation tags, and the unlabeled medical image samples
- the medical image sample is a medical image sample that has not been labeled
- the fourth processing unit is configured to input the labeled medical image sample and the unlabeled medical image sample into the segmentation model to be trained, and obtain the predicted segmentation result of each medical image sample by the segmentation model to be trained;
- the fifth processing unit is configured to input the predicted segmentation result into the discriminant model to obtain the discriminant result and confidence map of the discriminant model, wherein the confidence map is output by the upsampling layer in the discriminant model, so The discrimination result is output by the convolutional neural network in the discrimination model;
- a calculation unit configured to calculate a loss value regarding the segmentation model to be trained and the discrimination model based on a preset loss function according to the discrimination result, the confidence map, and the predicted segmentation result;
- the training unit is configured to train the discriminant model and the segmentation model to be trained based on the loss value, until the obtained loss value meets the preset loss condition, then the training is completed and the trained segmentation model is obtained.
- the calculation unit specifically includes:
- the first calculation subunit is configured to calculate the information about the segmentation model according to the first predicted segmentation sub-result corresponding to the labeled medical image sample and the real segmentation label of the labeled medical image sample in the predicted segmentation result.
- a second calculation subunit configured to calculate a second loss value of the segmentation model according to the second predicted segmentation sub-result corresponding to the unlabeled medical image sample in the predicted segmentation result and the confidence map;
- the third calculation subunit is configured to calculate the third loss value of the discriminant model according to the predicted segmentation result
- the fourth calculation subunit is configured to calculate the loss value according to the first loss value, the second loss value, and the third loss value.
- the second calculation subunit is specifically configured to:
- the fusion result of the output of the previous layer of the second intermediate layer and the output of the at least two first hierarchical structures is the output of the previous layer of the second intermediate layer and the output of the second intermediate layer.
- FIG. 7 is a schematic structural diagram of a terminal device provided by an embodiment of this application.
- the terminal device 7 of this embodiment includes: at least one processor 70 (only one is shown in FIG. 7), a memory 71, and is stored in the above-mentioned memory 71 and can run on the above-mentioned at least one processor 70
- the processor 70 executes the computer program 72, the steps in any of the medical image segmentation method embodiments described above are implemented.
- the aforementioned terminal device 7 may be a computing device such as a server, a mobile phone, a wearable device, an augmented reality (AR)/virtual reality (VR) device, a desktop computer, a notebook, a desktop computer, and a palmtop computer.
- the terminal device may include, but is not limited to, a processor 70 and a memory 71.
- FIG. 7 is only an example of the terminal device 7 and does not constitute a limitation on the terminal device 7. It may include more or less components than shown in the figure, or a combination of certain components, or different components. , For example, can also include input devices, output devices, network access devices, and so on.
- the above-mentioned input device may include a keyboard, a touch panel, a fingerprint collection sensor (used to collect user fingerprint information and fingerprint orientation information), a microphone, a camera, etc., and an output device may include a display, a speaker, and the like.
- the processor 70 may be a central processing unit (Central Processing Unit, CPU), and the processor 70 may also be other general-purpose processors, digital signal processors (Digital Signal Processors, DSPs), and application specific integrated circuits (Application Specific Integrated Circuits). , ASIC), Field-Programmable Gate Array (FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc.
- the general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
- the foregoing memory 71 may be an internal storage unit of the foregoing terminal device 7 in some embodiments, such as a hard disk or memory of the terminal device 7.
- the above-mentioned memory 71 may also be an external storage device of the above-mentioned terminal device 7 in other embodiments, for example, a plug-in hard disk equipped on the above-mentioned terminal device 7, a smart memory card (Smart Media Card, SMC), and a secure digital (Secure Digital). ,SD) card, flash card (Flash Card), etc.
- the aforementioned memory 71 may also include both an internal storage unit of the aforementioned terminal device 7 and an external storage device.
- the above-mentioned memory 71 is used to store an operating system, an application program, a boot loader (Boot Loader), data, and other programs, for example, the program code of the above-mentioned computer program.
- the aforementioned memory 71 can also be used to temporarily store data that has been output or will be output.
- the embodiments of the present application also provide a computer-readable storage medium, and the above-mentioned computer-readable storage medium stores a computer program, and when the above-mentioned computer program is executed by a processor, the steps in each of the above-mentioned method embodiments can be realized.
- the embodiments of the present application provide a computer program product.
- the terminal device can realize the steps in the foregoing method embodiments when the terminal device is executed.
- the above integrated unit is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a computer readable storage medium.
- this application implements all or part of the processes in the above-mentioned embodiments and methods, which can be completed by instructing relevant hardware through a computer program.
- the above-mentioned computer program can be stored in a computer-readable storage medium. When executed by the processor, the steps of the foregoing method embodiments can be implemented.
- the above-mentioned computer program includes computer program code, and the above-mentioned computer program code may be in the form of source code, object code, executable file, or some intermediate forms.
- the above-mentioned computer-readable medium may at least include: any entity or device capable of carrying computer program code to the camera device/terminal device, recording medium, computer memory, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), electric carrier signal, telecommunications signal and software distribution medium.
- ROM read-only memory
- RAM random access memory
- electric carrier signal telecommunications signal and software distribution medium.
- U disk mobile hard disk, floppy disk or CD-ROM, etc.
- computer-readable media cannot be electrical carrier signals and telecommunication signals.
- the disclosed apparatus/network equipment and method may be implemented in other ways.
- the device/network device embodiments described above are merely illustrative, and the division of the modules or units mentioned above is only a logical function division. In actual implementation, there may be other division methods, for example, multiple units or components may be It can be combined or integrated into another system, or some features can be ignored or not implemented.
- the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may be in electrical, mechanical or other forms.
- the units described above as separate components may or may not be physically separate, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of the embodiments.
Landscapes
- Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (20)
- 一种医学图像分割方法,其特征在于,包括:获取待检测医学图像;将所述待检测医学图像输入已训练的分割模型,其中,所述分割模型包括编码器和解码器,所述编码器包括多个第一层级结构,所述多个第一层级结构包括第一输入层、至少一个第一中间层和第一输出层,所述解码器包括第二输入层、至少一个第二中间层和第二输出层,其中,任意第二中间层的输入包括所述第二中间层的前一层的输出与至少两个第一层级结构的输出的融合结果;通过所述分割模型对所述待检测医学图像进行处理,获得所述分割模型的输出结果,其中,所述输出结果包括关于所述待检测医学图像中的医学特征区域的分割结果。
- 如权利要求1所述的医学图像分割方法,其特征在于,所述分割模型包括权重获取模块,所述权重获取模块位于所述编码器和所述解码器之间;所述通过所述分割模型对所述待检测医学图像进行处理,获得所述分割模型的输出结果,包括:通过所述编码器对所述待检测医学图像进行第一处理,获得所述编码器输出的第一特征矩阵;将所述第一特征矩阵输入所述权重获取模块;通过所述权重获取模块对所述第一特征矩阵进行第二处理,获得所述权重获取模块输出的权重矩阵;将所述权重矩阵与所述第一特征矩阵进行融合,获得第二特征矩阵;基于所述第二特征矩阵,通过所述解码器进行第三处理,获得所述输出结果。
- 如权利要求2所述的医学图像分割方法,其特征在于,所述通过所述权重获取模块对所述第一特征矩阵进行第二处理,获得所述权重获取模块输出的权重矩阵,包括:对所述第一特征矩阵进行第一卷积处理,获得第三特征矩阵;对所述第一特征矩阵进行第二卷积处理,获得第四特征矩阵;将所述第三特征矩阵与第四特征矩阵相乘,获得第五特征矩阵;通过激活函数对所述第五特征矩阵进行激活,获得所述权重矩阵。
- 如权利要求2所述的医学图像分割方法,其特征在于,所述基于所述第二特征矩阵,通过所述解码器进行第三处理,获得所述输出结果,包括:将所述第二特征矩阵与所述第一输出层的前一层的输出进行融合,获得第六特征矩阵;将所述第六特征矩阵输入所述解码器;基于所述第六特征矩阵,通过所述解码器进行第三处理,获得所述输出结果。
- 如权利要求1所述的医学图像分割方法,其特征在于,在将所述待检测医学图像输入已训练的分割模型之前,还包括:通过判别模型对待训练的分割模型进行训练,直到训练完成,并获得已训练的分割模型,其中,所述判别模型的输入包括所述待训练的分割模型的至少部分输出,所述判别模型包括卷积神经网络和上采样层,所述卷积神经网络的输出为所述上采样层的输入。
- 如权利要求5所述的医学图像分割方法,其特征在于,所述通过判别模型对待训练的分割模型进行训练,直到训练完成,并获得已训练的分割模型,包括:获取医学图像样本,其中,所述医学图像样本包括标注医学图像样本和未标注医学图像样本,所述标注医学图像样本为标注有真实分割标签的医学图像样本,所述未标注医学图像样本为未进行标注的医学图像样本;将所述标注医学图像样本和未标注医学图像样本输入待训练的分割模型,获得所述待训练的分割模型对各个医学图像样本的预测分割结果;将所述预测分割结果输入所述判别模型,获得所述判别模型的判别结果和置信图,其中,所述置信图由所述判别模型中的上采样层输出,所述判别结果由所述判别模型中的卷积神经网络输出;根据所述判别结果、所述置信图和所述预测分割结果,基于预设损失函数计算关于所述待训练的分割模型和所述判别模型的损失值;基于所述损失值对所述判别模型和所述待训练的分割模型进行训练,直到所得到的损失值符合预设损失条件,则训练完成,并获得已训练的分割模型。
- 如权利要求6所述的医学图像分割方法,其特征在于,所述根据所述判别结果、所述置信图和所述预测分割结果,基于预设损失函数计算关于所述待训练的分割模型和所述判别模型的损失值,包括:根据所述预测分割结果中,所述标注医学图像样本所对应的第一预测分割子结果与所述标注医学图像样本的真实分割标签,计算关于所述分割模型的第一损失值;根据所述预测分割结果中所述未标注医学图像样本所对应的第二预测分割子结果和所述置信图,计算所述分割模型的第二损失值;根据所述预测分割结果,计算关于所述判别模型的第三损失值;根据所述第一损失值、第二损失值和第三损失值,计算得到所述损失值。
- 如权利要求7所述的医学图像分割方法,其特征在于,所述根据所述预测分割 结果中所述未标注医学图像样本所对应的第二预测分割子结果和所述置信图,计算所述分割模型的第二损失值,包括:对所述置信图中的各个位置进行编码运算,获得所述置信图所对应的编码图像,所述编码图像包含所述置信图中的每个位置的编码值;根据编码图像和所述未标注医学图像样本所对应的第二预测分割子结果,计算所述分割模型的第二损失值。
- 如权利要求1至8任意一项所述的医学图像分割方法,其特征在于,所述第二中间层的前一层的输出与至少两个第一层级结构的输出的融合结果为所述第二中间层的前一层的输出、与所述第二中间层相对应的第一层级结构的输出以及与所述第二中间层相对应的第一层级结构的前一层的输出的融合结果。
- 一种医学图像分割装置,其特征在于,包括:第一获取模块,用于获取待检测医学图像;输入模块,用于将所述待检测医学图像输入已训练的分割模型,其中,所述分割模型包括编码器和解码器,所述编码器包括多个第一层级结构,所述多个第一层级结构包括第一输入层、至少一个第一中间层和第一输出层,所述解码器包括第二输入层、至少一个第二中间层和第二输出层,其中,任意第二中间层的输入包括所述第二中间层的前一层的输出与至少两个第一层级结构的输出的融合结果;处理模块,用于通过所述分割模型对所述待检测医学图像进行处理,获得所述分割模型的输出结果,其中,所述输出结果包括关于所述待检测医学图像中的医学特征区域的分割结果。
- 一种终端设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,其特征在于,所述处理器执行所述计算机程序时实现如下步骤:获取待检测医学图像;将所述待检测医学图像输入已训练的分割模型,其中,所述分割模型包括编码器和解码器,所述编码器包括多个第一层级结构,所述多个第一层级结构包括第一输入层、至少一个第一中间层和第一输出层,所述解码器包括第二输入层、至少一个第二中间层和第二输出层,其中,任意第二中间层的输入包括所述第二中间层的前一层的输出与至少两个第一层级结构的输出的融合结果;通过所述分割模型对所述待检测医学图像进行处理,获得所述分割模型的输出结果,其中,所述输出结果包括关于所述待检测医学图像中的医学特征区域的分割结果。
- 如权利要求11所述的终端设备,其特征在于,所述分割模型包括权重获取模 块,所述权重获取模块位于所述编码器和所述解码器之间;所述处理器执行所述计算机程序时,所述通过所述分割模型对所述待检测医学图像进行处理,获得所述分割模型的输出结果,包括:通过所述编码器对所述待检测医学图像进行第一处理,获得所述编码器输出的第一特征矩阵;将所述第一特征矩阵输入所述权重获取模块;通过所述权重获取模块对所述第一特征矩阵进行第二处理,获得所述权重获取模块输出的权重矩阵;将所述权重矩阵与所述第一特征矩阵进行融合,获得第二特征矩阵;基于所述第二特征矩阵,通过所述解码器进行第三处理,获得所述输出结果。
- 如权利要求12所述的终端设备,其特征在于,所述处理器执行所述计算机程序时,所述通过所述权重获取模块对所述第一特征矩阵进行第二处理,获得所述权重获取模块输出的权重矩阵,包括:对所述第一特征矩阵进行第一卷积处理,获得第三特征矩阵;对所述第一特征矩阵进行第二卷积处理,获得第四特征矩阵;将所述第三特征矩阵与第四特征矩阵相乘,获得第五特征矩阵;通过激活函数对所述第五特征矩阵进行激活,获得所述权重矩阵。
- 如权利要求12所述的终端设备,其特征在于,所述处理器执行所述计算机程序时,所述基于所述第二特征矩阵,通过所述解码器进行第三处理,获得所述输出结果,包括:将所述第二特征矩阵与所述第一输出层的前一层的输出进行融合,获得第六特征矩阵;将所述第六特征矩阵输入所述解码器;基于所述第六特征矩阵,通过所述解码器进行第三处理,获得所述输出结果。
- 如权利要求11所述的终端设备,其特征在于,所述处理器执行所述计算机程序时,在将所述待检测医学图像输入已训练的分割模型之前,还包括:通过判别模型对待训练的分割模型进行训练,直到训练完成,并获得已训练的分割模型,其中,所述判别模型的输入包括所述待训练的分割模型的至少部分输出,所述判别模型包括卷积神经网络和上采样层,所述卷积神经网络的输出为所述上采样层的输入。
- 如权利要求15所述的终端设备,其特征在于,所述处理器执行所述计算机程序时,所述通过判别模型对待训练的分割模型进行训练,直到训练完成,并获得已训 练的分割模型,包括:获取医学图像样本,其中,所述医学图像样本包括标注医学图像样本和未标注医学图像样本,所述标注医学图像样本为标注有真实分割标签的医学图像样本,所述未标注医学图像样本为未进行标注的医学图像样本;将所述标注医学图像样本和未标注医学图像样本输入待训练的分割模型,获得所述待训练的分割模型对各个医学图像样本的预测分割结果;将所述预测分割结果输入所述判别模型,获得所述判别模型的判别结果和置信图,其中,所述置信图由所述判别模型中的上采样层输出,所述判别结果由所述判别模型中的卷积神经网络输出;根据所述判别结果、所述置信图和所述预测分割结果,基于预设损失函数计算关于所述待训练的分割模型和所述判别模型的损失值;基于所述损失值对所述判别模型和所述待训练的分割模型进行训练,直到所得到的损失值符合预设损失条件,则训练完成,并获得已训练的分割模型。
- 如权利要求16所述的终端设备,其特征在于,所述处理器执行所述计算机程序时,所述根据所述判别结果、所述置信图和所述预测分割结果,基于预设损失函数计算关于所述待训练的分割模型和所述判别模型的损失值,包括:根据所述预测分割结果中,所述标注医学图像样本所对应的第一预测分割子结果与所述标注医学图像样本的真实分割标签,计算关于所述分割模型的第一损失值;根据所述预测分割结果中所述未标注医学图像样本所对应的第二预测分割子结果和所述置信图,计算所述分割模型的第二损失值;根据所述预测分割结果,计算关于所述判别模型的第三损失值;根据所述第一损失值、第二损失值和第三损失值,计算得到所述损失值。
- 如权利要求17所述的终端设备,其特征在于,所述处理器执行所述计算机程序时,所述根据所述预测分割结果中所述未标注医学图像样本所对应的第二预测分割子结果和所述置信图,计算所述分割模型的第二损失值,包括:对所述置信图中的各个位置进行编码运算,获得所述置信图所对应的编码图像,所述编码图像包含所述置信图中的每个位置的编码值;根据编码图像和所述未标注医学图像样本所对应的第二预测分割子结果,计算所述分割模型的第二损失值。
- 如权利要求11至18任意一项所述的终端设备,其特征在于,所述处理器执行所述计算机程序时,所述第二中间层的前一层的输出与至少两个第一层级结构的输出的融合结果为所述第二中间层的前一层的输出、与所述第二中间层相对应的第一层级 结构的输出以及与所述第二中间层相对应的第一层级结构的前一层的输出的融合结果。
- 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现如权利要求1至9任一项所述的医学图像分割方法。
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2020/078800 WO2021179205A1 (zh) | 2020-03-11 | 2020-03-11 | 医学图像分割方法、医学图像分割装置及终端设备 |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2020/078800 WO2021179205A1 (zh) | 2020-03-11 | 2020-03-11 | 医学图像分割方法、医学图像分割装置及终端设备 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021179205A1 true WO2021179205A1 (zh) | 2021-09-16 |
Family
ID=77671674
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/078800 WO2021179205A1 (zh) | 2020-03-11 | 2020-03-11 | 医学图像分割方法、医学图像分割装置及终端设备 |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2021179205A1 (zh) |
Cited By (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113850826A (zh) * | 2021-09-27 | 2021-12-28 | 平安科技(深圳)有限公司 | 基于图像分割的心脏图像处理方法、装置、设备及介质 |
CN113902827A (zh) * | 2021-12-02 | 2022-01-07 | 北京鹰瞳科技发展股份有限公司 | 一种针对皮肤病的愈后效果预测***、方法以及电子设备 |
CN114066913A (zh) * | 2022-01-12 | 2022-02-18 | 广东工业大学 | 一种心脏图像分割方法及*** |
CN114724133A (zh) * | 2022-04-18 | 2022-07-08 | 北京百度网讯科技有限公司 | 文字检测和模型训练方法、装置、设备及存储介质 |
US20220230321A1 (en) * | 2021-01-15 | 2022-07-21 | Adobe Inc. | Generating class-agnostic object masks in digital images |
CN114912575A (zh) * | 2022-04-06 | 2022-08-16 | 西安交通大学 | 基于连接Swin Transformer通路的医学图像分割模型及方法 |
CN114972293A (zh) * | 2022-06-14 | 2022-08-30 | 深圳市大数据研究院 | 基于半监督时空注意力网络的视频息肉分割方法、装置 |
CN114972297A (zh) * | 2022-06-15 | 2022-08-30 | 北京理工大学 | 口腔健康监测方法及装置 |
CN115147526A (zh) * | 2022-06-30 | 2022-10-04 | 北京百度网讯科技有限公司 | 服饰生成模型的训练、生成服饰图像的方法和装置 |
CN115546239A (zh) * | 2022-11-30 | 2022-12-30 | 珠海横琴圣澳云智科技有限公司 | 基于边界注意力与距离变换的目标分割方法和装置 |
CN115661449A (zh) * | 2022-09-22 | 2023-01-31 | 北京百度网讯科技有限公司 | 图像分割与图像分割模型的训练方法、装置 |
CN115713535A (zh) * | 2022-11-07 | 2023-02-24 | 阿里巴巴(中国)有限公司 | 图像分割模型确定方法以及图像分割方法 |
US11631234B2 (en) | 2019-07-22 | 2023-04-18 | Adobe, Inc. | Automatically detecting user-requested objects in images |
CN116188879A (zh) * | 2023-04-27 | 2023-05-30 | 广州医思信息科技有限公司 | 图像分类、图像分类模型训练方法、装置、设备及介质 |
US11681919B2 (en) | 2020-03-12 | 2023-06-20 | Adobe Inc. | Automatically selecting query objects in digital images |
CN116503420A (zh) * | 2023-04-26 | 2023-07-28 | 佛山科学技术学院 | 一种基于联邦学习的图像分割方法及相关设备 |
WO2023165033A1 (zh) * | 2022-03-02 | 2023-09-07 | 深圳硅基智能科技有限公司 | 识别医学图像中的目标的模型训练、方法、设备及介质 |
CN116758093A (zh) * | 2023-05-30 | 2023-09-15 | 首都医科大学宣武医院 | 一种图像分割方法、模型训练方法、装置、设备及介质 |
US11797847B2 (en) | 2019-07-22 | 2023-10-24 | Adobe Inc. | Selecting instances of detected objects in images utilizing object detection models |
CN116993762A (zh) * | 2023-09-26 | 2023-11-03 | 腾讯科技(深圳)有限公司 | 图像分割方法、装置、电子设备及存储介质 |
CN117115444A (zh) * | 2023-09-08 | 2023-11-24 | 北京卓视智通科技有限责任公司 | 一种多任务图像分割方法、***、计算机设备和存储介质 |
US11886494B2 (en) | 2020-02-25 | 2024-01-30 | Adobe Inc. | Utilizing natural language processing automatically select objects in images |
CN117853507A (zh) * | 2024-03-06 | 2024-04-09 | 阿里巴巴(中国)有限公司 | 交互式图像分割方法、设备、存储介质和程序产品 |
US11972569B2 (en) | 2021-01-26 | 2024-04-30 | Adobe Inc. | Segmenting objects in digital images utilizing a multi-object segmentation model framework |
US12020414B2 (en) | 2019-07-22 | 2024-06-25 | Adobe Inc. | Utilizing deep neural networks to automatically select instances of detected objects in images |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108876792A (zh) * | 2018-04-13 | 2018-11-23 | 北京迈格威科技有限公司 | 语义分割方法、装置和***及存储介质 |
CN109034162A (zh) * | 2018-07-13 | 2018-12-18 | 南京邮电大学 | 一种图像语义分割方法 |
CN109784380A (zh) * | 2018-12-27 | 2019-05-21 | 西安交通大学 | 一种基于生成对抗学习的多维度田间杂草识别方法 |
CN110197493A (zh) * | 2019-05-24 | 2019-09-03 | 清华大学深圳研究生院 | 眼底图像血管分割方法 |
US20190333198A1 (en) * | 2018-04-25 | 2019-10-31 | Adobe Inc. | Training and utilizing an image exposure transformation neural network to generate a long-exposure image from a single short-exposure image |
CN110503654A (zh) * | 2019-08-01 | 2019-11-26 | 中国科学院深圳先进技术研究院 | 一种基于生成对抗网络的医学图像分割方法、***及电子设备 |
CN110675406A (zh) * | 2019-09-16 | 2020-01-10 | 南京信息工程大学 | 基于残差双注意力深度网络的ct图像肾脏分割算法 |
CN110689083A (zh) * | 2019-09-30 | 2020-01-14 | 苏州大学 | 一种上下文金字塔融合网络及图像分割方法 |
CN110807762A (zh) * | 2019-09-19 | 2020-02-18 | 温州大学 | 一种基于gan的视网膜血管图像智能分割方法 |
-
2020
- 2020-03-11 WO PCT/CN2020/078800 patent/WO2021179205A1/zh active Application Filing
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108876792A (zh) * | 2018-04-13 | 2018-11-23 | 北京迈格威科技有限公司 | 语义分割方法、装置和***及存储介质 |
US20190333198A1 (en) * | 2018-04-25 | 2019-10-31 | Adobe Inc. | Training and utilizing an image exposure transformation neural network to generate a long-exposure image from a single short-exposure image |
CN109034162A (zh) * | 2018-07-13 | 2018-12-18 | 南京邮电大学 | 一种图像语义分割方法 |
CN109784380A (zh) * | 2018-12-27 | 2019-05-21 | 西安交通大学 | 一种基于生成对抗学习的多维度田间杂草识别方法 |
CN110197493A (zh) * | 2019-05-24 | 2019-09-03 | 清华大学深圳研究生院 | 眼底图像血管分割方法 |
CN110503654A (zh) * | 2019-08-01 | 2019-11-26 | 中国科学院深圳先进技术研究院 | 一种基于生成对抗网络的医学图像分割方法、***及电子设备 |
CN110675406A (zh) * | 2019-09-16 | 2020-01-10 | 南京信息工程大学 | 基于残差双注意力深度网络的ct图像肾脏分割算法 |
CN110807762A (zh) * | 2019-09-19 | 2020-02-18 | 温州大学 | 一种基于gan的视网膜血管图像智能分割方法 |
CN110689083A (zh) * | 2019-09-30 | 2020-01-14 | 苏州大学 | 一种上下文金字塔融合网络及图像分割方法 |
Non-Patent Citations (1)
Title |
---|
JIAWEI ZHANG , YUZHEN JIN , JILAN XU , XIAOWEI XU , YANCHUN ZHANG: "MDU-Net: Multi-scale Densely Connected U-Net for biomedical image segmentation", ARXIV.ORG, 2 December 2018 (2018-12-02), pages 1 - 10, XP080987964 * |
Cited By (40)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11631234B2 (en) | 2019-07-22 | 2023-04-18 | Adobe, Inc. | Automatically detecting user-requested objects in images |
US11797847B2 (en) | 2019-07-22 | 2023-10-24 | Adobe Inc. | Selecting instances of detected objects in images utilizing object detection models |
US12020414B2 (en) | 2019-07-22 | 2024-06-25 | Adobe Inc. | Utilizing deep neural networks to automatically select instances of detected objects in images |
US11886494B2 (en) | 2020-02-25 | 2024-01-30 | Adobe Inc. | Utilizing natural language processing automatically select objects in images |
US11681919B2 (en) | 2020-03-12 | 2023-06-20 | Adobe Inc. | Automatically selecting query objects in digital images |
US11587234B2 (en) * | 2021-01-15 | 2023-02-21 | Adobe Inc. | Generating class-agnostic object masks in digital images |
US11900611B2 (en) | 2021-01-15 | 2024-02-13 | Adobe Inc. | Generating object masks of object parts utlizing deep learning |
US20220230321A1 (en) * | 2021-01-15 | 2022-07-21 | Adobe Inc. | Generating class-agnostic object masks in digital images |
US20230136913A1 (en) * | 2021-01-15 | 2023-05-04 | Adobe Inc. | Generating object masks of object parts utlizing deep learning |
US11972569B2 (en) | 2021-01-26 | 2024-04-30 | Adobe Inc. | Segmenting objects in digital images utilizing a multi-object segmentation model framework |
CN113850826A (zh) * | 2021-09-27 | 2021-12-28 | 平安科技(深圳)有限公司 | 基于图像分割的心脏图像处理方法、装置、设备及介质 |
CN113902827B (zh) * | 2021-12-02 | 2022-03-22 | 北京鹰瞳科技发展股份有限公司 | 一种针对皮肤病的愈后效果预测***、方法以及电子设备 |
CN113902827A (zh) * | 2021-12-02 | 2022-01-07 | 北京鹰瞳科技发展股份有限公司 | 一种针对皮肤病的愈后效果预测***、方法以及电子设备 |
CN114066913A (zh) * | 2022-01-12 | 2022-02-18 | 广东工业大学 | 一种心脏图像分割方法及*** |
WO2023165033A1 (zh) * | 2022-03-02 | 2023-09-07 | 深圳硅基智能科技有限公司 | 识别医学图像中的目标的模型训练、方法、设备及介质 |
CN114912575A (zh) * | 2022-04-06 | 2022-08-16 | 西安交通大学 | 基于连接Swin Transformer通路的医学图像分割模型及方法 |
CN114912575B (zh) * | 2022-04-06 | 2024-04-09 | 西安交通大学 | 基于连接Swin Transformer通路的医学图像分割模型及方法 |
CN114724133B (zh) * | 2022-04-18 | 2024-02-02 | 北京百度网讯科技有限公司 | 文字检测和模型训练方法、装置、设备及存储介质 |
CN114724133A (zh) * | 2022-04-18 | 2022-07-08 | 北京百度网讯科技有限公司 | 文字检测和模型训练方法、装置、设备及存储介质 |
CN114972293A (zh) * | 2022-06-14 | 2022-08-30 | 深圳市大数据研究院 | 基于半监督时空注意力网络的视频息肉分割方法、装置 |
CN114972297A (zh) * | 2022-06-15 | 2022-08-30 | 北京理工大学 | 口腔健康监测方法及装置 |
CN115147526A (zh) * | 2022-06-30 | 2022-10-04 | 北京百度网讯科技有限公司 | 服饰生成模型的训练、生成服饰图像的方法和装置 |
CN115147526B (zh) * | 2022-06-30 | 2023-09-26 | 北京百度网讯科技有限公司 | 服饰生成模型的训练、生成服饰图像的方法和装置 |
CN115661449A (zh) * | 2022-09-22 | 2023-01-31 | 北京百度网讯科技有限公司 | 图像分割与图像分割模型的训练方法、装置 |
CN115661449B (zh) * | 2022-09-22 | 2023-11-21 | 北京百度网讯科技有限公司 | 图像分割与图像分割模型的训练方法、装置 |
CN115713535A (zh) * | 2022-11-07 | 2023-02-24 | 阿里巴巴(中国)有限公司 | 图像分割模型确定方法以及图像分割方法 |
CN115713535B (zh) * | 2022-11-07 | 2024-05-14 | 阿里巴巴(中国)有限公司 | 图像分割模型确定方法以及图像分割方法 |
CN115546239A (zh) * | 2022-11-30 | 2022-12-30 | 珠海横琴圣澳云智科技有限公司 | 基于边界注意力与距离变换的目标分割方法和装置 |
CN115546239B (zh) * | 2022-11-30 | 2023-04-07 | 珠海横琴圣澳云智科技有限公司 | 基于边界注意力与距离变换的目标分割方法和装置 |
CN116503420B (zh) * | 2023-04-26 | 2024-05-14 | 佛山科学技术学院 | 一种基于联邦学习的图像分割方法及相关设备 |
CN116503420A (zh) * | 2023-04-26 | 2023-07-28 | 佛山科学技术学院 | 一种基于联邦学习的图像分割方法及相关设备 |
CN116188879A (zh) * | 2023-04-27 | 2023-05-30 | 广州医思信息科技有限公司 | 图像分类、图像分类模型训练方法、装置、设备及介质 |
CN116188879B (zh) * | 2023-04-27 | 2023-11-28 | 广州医思信息科技有限公司 | 图像分类、图像分类模型训练方法、装置、设备及介质 |
CN116758093A (zh) * | 2023-05-30 | 2023-09-15 | 首都医科大学宣武医院 | 一种图像分割方法、模型训练方法、装置、设备及介质 |
CN116758093B (zh) * | 2023-05-30 | 2024-05-07 | 首都医科大学宣武医院 | 一种图像分割方法、模型训练方法、装置、设备及介质 |
CN117115444A (zh) * | 2023-09-08 | 2023-11-24 | 北京卓视智通科技有限责任公司 | 一种多任务图像分割方法、***、计算机设备和存储介质 |
CN117115444B (zh) * | 2023-09-08 | 2024-04-16 | 北京卓视智通科技有限责任公司 | 一种多任务图像分割方法、***、计算机设备和存储介质 |
CN116993762B (zh) * | 2023-09-26 | 2024-01-19 | 腾讯科技(深圳)有限公司 | 图像分割方法、装置、电子设备及存储介质 |
CN116993762A (zh) * | 2023-09-26 | 2023-11-03 | 腾讯科技(深圳)有限公司 | 图像分割方法、装置、电子设备及存储介质 |
CN117853507A (zh) * | 2024-03-06 | 2024-04-09 | 阿里巴巴(中国)有限公司 | 交互式图像分割方法、设备、存储介质和程序产品 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021179205A1 (zh) | 医学图像分割方法、医学图像分割装置及终端设备 | |
CN111429464B (zh) | 医学图像分割方法、医学图像分割装置及终端设备 | |
CN112992308B (zh) | 医学图像报告生成模型的训练方法及图像报告生成方法 | |
WO2020215974A1 (zh) | 用于人体检测的方法和装置 | |
Borsting et al. | Applied deep learning in plastic surgery: classifying rhinoplasty with a mobile app | |
EP4266195A1 (en) | Training of text and image models | |
CN113077434B (zh) | 基于多模态信息的肺癌识别方法、装置及存储介质 | |
CN112712879A (zh) | 医学影像报告的信息提取方法、装置、设备及存储介质 | |
CN116304307A (zh) | 一种图文跨模态检索网络训练方法、应用方法及电子设备 | |
Kaur et al. | Methods for automatic generation of radiological reports of chest radiographs: a comprehensive survey | |
Liang et al. | Dense networks with relative location awareness for thorax disease identification | |
Chen et al. | HADCNet: Automatic segmentation of COVID-19 infection based on a hybrid attention dense connected network with dilated convolution | |
Chung et al. | Prediction of oxygen requirement in patients with COVID-19 using a pre-trained chest radiograph xAI model: efficient development of auditable risk prediction models via a fine-tuning approach | |
Liang et al. | FCF: Feature complement fusion network for detecting COVID-19 through CT scan images | |
CN117112829B (zh) | 医疗数据跨模态检索方法、装置和相关设备 | |
Gaggion et al. | CheXmask: a large-scale dataset of anatomical segmentation masks for multi-center chest x-ray images | |
CN113724185B (zh) | 用于图像分类的模型处理方法、装置及存储介质 | |
Takagi et al. | Transformer-based personalized attention mechanism for medical images with clinical records | |
Xu et al. | Hybrid reinforced medical report generation with m-linear attention and repetition penalty | |
Shetty et al. | Multimodal medical tensor fusion network-based DL framework for abnormality prediction from the radiology CXRs and clinical text reports | |
CN113159053A (zh) | 图像识别方法、装置及计算设备 | |
Song et al. | Multi-scale Superpixel based Hierarchical Attention model for brain CT classification | |
CN114724016A (zh) | 图像分类方法、计算机设备和存储介质 | |
Pang et al. | Correlation matters: multi-scale fine-grained contextual information extraction for hepatic tumor segmentation | |
Ni et al. | FSE-Net: Rethinking the up-sampling operation in encoder-decoder structure for retinal vessel segmentation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20924952 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20924952 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20924952 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 10.07.2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20924952 Country of ref document: EP Kind code of ref document: A1 |