CN112686898B - Automatic radiotherapy target area segmentation method based on self-supervision learning - Google Patents

Automatic radiotherapy target area segmentation method based on self-supervision learning Download PDF

Info

Publication number
CN112686898B
CN112686898B CN202110274005.3A CN202110274005A CN112686898B CN 112686898 B CN112686898 B CN 112686898B CN 202110274005 A CN202110274005 A CN 202110274005A CN 112686898 B CN112686898 B CN 112686898B
Authority
CN
China
Prior art keywords
segmentation
training
data
network
model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110274005.3A
Other languages
Chinese (zh)
Other versions
CN112686898A (en
Inventor
章毅
柏森
余程嵘
宋�莹
胡俊杰
王强
张海仙
郭际香
郭泉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sichuan University
Original Assignee
Sichuan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sichuan University filed Critical Sichuan University
Priority to CN202110274005.3A priority Critical patent/CN112686898B/en
Publication of CN112686898A publication Critical patent/CN112686898A/en
Application granted granted Critical
Publication of CN112686898B publication Critical patent/CN112686898B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses a radiotherapy target area automatic segmentation method based on self-supervision learning, which relates to the technical field of image processing and comprises the following steps: 1) preparing data: collecting original CT data, separating a labeled data set and a non-labeled data set, and sketching the labeled data set; 2) feature extraction: constructing a pre-training network based on self-supervision learning, inputting a label-free data set into the pre-training network for iterative training, and selecting an optimal pre-training model; 3) and (3) generating a segmentation model: and constructing a segmentation network, loading the trained self-supervision pre-training model into the segmentation network, inputting the labeled data set into the segmentation network for iterative training to select an optimal model, and finally testing and evaluating the segmentation performance of the model. The invention uses the coordinate label of CT data to pre-train the self-supervision task without designing a new label; the pre-trained model contains shallow features of the CT image, and therefore has a faster convergence rate when the segmentation task is performed.

Description

Automatic radiotherapy target area segmentation method based on self-supervision learning
Technical Field
The invention relates to the technical field of image processing, in particular to the technical field of an automatic segmentation method of a radiotherapy target area based on self-supervision learning.
Background
Radiotherapy (abbreviated as radiotherapy) is one of the important means for treating malignant tumors at present, and the purpose of treatment is mainly achieved by irradiating tumor parts with a certain dose of radioactive rays. Modern radiotherapy comprises a series of normative procedures, and the main work of the radiotherapy comprises target region delineation, plan making, plan implementation, reverse optimization and the like. Wherein the target area is sketched mainly by the physical engineer to CT image data carry out the sketching of target area and organs at risk, and the time proportion that the sketching task of common case is all great, and the difference on the result is often brought because of the difference of the custom of sketching to artificial sketching.
The method aims to reduce the time cost of the sketching task in the radiotherapy process and further reduce the influence of the artificial sketching habit on the sketching result. An image processing algorithm is often introduced into the delineation task to realize automatic delineation, wherein a segmentation algorithm based on a deep neural network is one of important means for realizing medical image processing. The neural network is mainly formed by connecting and combining a plurality of neurons and mainly comprises an input layer, a hidden layer and an output layer. Weight information is carried among the neurons of the hidden layer, and the weight is continuously updated to approximate the highly nonlinear relation between the input data and the characteristics of the input data through the previous calculation and the backward propagation during training. After multiple times of training, the network output is converged, an optimal network model is selected according to the evaluation index, and finally, the input image can be automatically segmented based on the model.
In the prior art, a deep neural network-based automatic delineation is adopted, firstly, a physical engineer needs to delineate a large amount of CT image data to be trained, the delineated data are divided into a training set and a test set according to a certain proportion, a network structure is designed, the training set data are input into a network for training, an optimal model is selected according to the test set after training is iterated for multiple times, and therefore the final model is determined to be put into use.
However, the prior art has the following disadvantages: in the current automatic delineation technology based on the deep neural network, in order to achieve an optimal delineation model, a large amount of training data is required to be input in the early stage, which increases the time cost and equipment cost brought by the training model in the early stage work.
Disclosure of Invention
The invention aims to: in order to solve the technical problem, the invention provides a radiotherapy target region automatic segmentation method based on self-supervision learning.
The invention specifically adopts the following technical scheme for realizing the purpose:
a radiotherapy target area automatic segmentation method based on self-supervision learning comprises the following steps:
step 1, data preparation: collecting original CT data, separating a labeled data set and a non-labeled data set, and sketching the labeled data set;
step 2, feature extraction: constructing a pre-training network based on self-supervision learning, inputting a label-free data set into the pre-training network for iterative training, and selecting an optimal pre-training model;
step 3, generating a segmentation model: and constructing a segmentation network, loading the trained self-supervision pre-training model into the segmentation network, inputting the labeled data set into the segmentation network for iterative training to select an optimal model, and finally testing and evaluating the segmentation performance of the model.
Further, in step 1, the data preparation specifically includes the following steps:
step 1a, data acquisition: a physicist scans a target region of a patient by using a CT machine to obtain CT image data, and a plurality of continuous CT images are generated by one-time scanning;
step 1b, data division: dividing the obtained original CT data into a non-labeled data set and a labeled data set according to the ratio of 6:4, and respectively corresponding to a pre-learning task and a segmentation task;
step 1c, data processing: and (4) performing target region delineation on the labeled data set in the step 1b by a physicist according to pathological features of the case, and generating a data set containing target region segmentation labels.
Further, in step 2, the feature extraction mainly generates a pre-training model through an executed pre-learning task, and specifically includes the following steps:
step 2a, constructing a pre-learning network: the pre-learning task inputs two different CT images belonging to a case each time, and the output result is a single numerical value each time, so that the pre-learning network is realized by combining two twin deep neural networks in a weight sharing mode, and the pre-learning network comprises two modules of a coder and a fusion device;
step 2b, sample data set construction: the CT data acquired in step 1 contains a plurality of cases, and it is assumed that a patient is waiting for treatmentThe physical CT data comprises N cases, each case comprises M CT images, and the size of the constructed sample set is equal to that of the sample set
Figure 733578DEST_PATH_IMAGE001
Each sample can be represented as a triplet slice1, slice2, Dis, where slice1 and slice2 represent two CT images of the same case, respectively, and Dis represents the relative distance between the two CT images, which is calculated as:
Figure 244194DEST_PATH_IMAGE002
wherein the content of the first and second substances,
Figure 492772DEST_PATH_IMAGE003
are the coordinate values of the CT image,
Figure 622402DEST_PATH_IMAGE004
is the relative distance between the two CT images;
step 2c, CT data preprocessing, namely performing data augmentation operation on the unlabeled CT image data obtained in the step 1, and realizing the data augmentation operation through random overturning, translation and rotation, so that the pre-learning model can learn the characteristics with stronger robustness;
step 2d, pre-learning model training: and (4) the augmented CT data is calculated according to the following weight ratio of 8: and 2, dividing the proportion of the data into a training set and a test set, inputting the data of the training set into the pre-learning network designed in the step 2a, sequentially performing previous calculation, back propagation and repetition for multiple times until the output loss value is converged, and stopping training when the output on the test set is the minimum value in multiple times of training, thereby completing feature extraction and obtaining the pre-learning model.
Further, in step 2a, the encoder mainly comprises a plurality of convolution layers and pooling layers, and respectively extracts the features of two input CT images to obtain two decoded feature vectors;
the fusion device aims to calculate and fuse the feature vectors extracted by the two networks, the method adopted by the fusion is a vector maximum interpolation method, namely, the maximum value on each channel in the two vectors is selected as a fusion vector, and the calculation formula is as follows:
Figure 46562DEST_PATH_IMAGE005
wherein
Figure 337866DEST_PATH_IMAGE006
Represents the output vector after the two networks are merged,
Figure 932795DEST_PATH_IMAGE007
Figure 600537DEST_PATH_IMAGE008
respectively representing the output vectors of the twin deep neural network 1 and the twin deep neural network 2.
Further, in step 3, a pre-training model is obtained after feature extraction is completed, the pre-training model carries shallow image features of CT data, and in this step, a small amount of data containing segmentation labels are input into a segmentation network for training based on the pre-training model to obtain a segmentation model, and the specific steps are as follows:
step 3a, constructing a segmentation network, wherein the segmentation model network is designed based on a deep neural network and comprises an encoder, a random multi-scale module and a decoder;
step 3b, loading a pre-learning model: because the structure of the encoder of the pre-learning network is consistent with that of the encoder of the segmentation network during network design, the pre-training model obtained in the step 2 is directly loaded into the segmentation network, so that the segmentation network can directly obtain the characteristics of the CT image and can be suitable for scenes with too little label data;
step 3c, CT data preprocessing: considering that the labeled CT data volume is small, in order to make the segmentation model effect more feasible, the labeled CT data set obtained in step 1 needs to be subjected to augmentation operation, which is mainly implemented by random inversion, translation and rotation;
step 3d, training a segmentation model: the method comprises the following steps of performing model training by using a deep neural network for image segmentation; firstly, the data after the augmentation operation is divided into 7: 1: 2, dividing the proportion of the training set, the verification set and the test set, inputting the training set data into the network designed in the step 3a, sequentially performing previous item calculation, back propagation and repetition for multiple times until the loss value output by the network is converged, and stopping training when the output on the verification set is the minimum value in multiple times of training, thereby obtaining the optimal model result of the segmentation model network;
step 3e, segmentation model testing: after the training of the segmentation network model is completed, the segmentation effect of the segmentation model on the test set in (3d) needs to be quantitatively evaluated, and commonly used evaluation indexes include TPVF, PPV, and DSC, which are defined as follows:
Figure 613623DEST_PATH_IMAGE009
Figure 75829DEST_PATH_IMAGE010
Figure 726032DEST_PATH_IMAGE011
wherein
Figure 790940DEST_PATH_IMAGE012
And
Figure 127375DEST_PATH_IMAGE013
the model is used for evaluating the segmentation result, and the model is used for evaluating the segmentation result and is mainly used for evaluating the TPVF and the PPV.
Further, in step 3a, the encoder is consistent with the structure of the encoder of the pre-training network, and the segmentation network encoder is composed of a series of convolution and pooling layers and is used for extracting abstract features from the input CT image.
Further, in step 3a, the random multi-scale module mainly comprises a cavity convolution module with 4 different cavity rates and a global average pooling layer, and the random multi-scale module further processes the feature vector output by the encoder to enable the feature vector to have features with different scales, so as to improve the robustness of the model to the features with different scales.
Further, in step 3a, the decoder mainly aims to restore the previous features to the original input size, generate a class prediction for the original picture on each pixel, and merge the deep features and the shallow features, so that the feature multiplexing of the model can be improved, and the convergence of the model can be accelerated.
The invention has the following beneficial effects:
1. the invention uses the coordinate label of CT data to pre-train the self-supervision task without designing a new label; the pre-training model generated in the pre-learning task contains shallow features of the CT image, so that the convergence speed is high when the segmentation task is executed; since the pre-trained model contains the general features of CT data, segmentation model training can also be performed for most organs at risk and target regions.
2. The traditional target region segmentation method based on the deep neural network needs a large amount of labeled CT data, the actual labeled CT data are extremely few, and a physical engineer is required to consume a great amount of time and cost to draw up training data. The method can train the segmentation model only by less labeled data, greatly saves the labor and time cost, and has important guiding significance for the automatic segmentation work of the target area based on the neural network in the future.
Drawings
FIG. 1 is a schematic flow diagram of the system of the present invention;
FIG. 2 is a schematic diagram of "relative position" of a CT scan;
FIG. 3 is a schematic diagram of a pre-learning network architecture;
fig. 4 is a schematic diagram of a split network architecture.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, but not all, embodiments of the present invention. The components of embodiments of the present invention generally described and illustrated in the figures herein may be arranged and designed in a wide variety of different configurations.
Example 1
As shown in fig. 1, the present embodiment provides a method for automatically segmenting a radiotherapy target region based on self-supervised learning, including the following steps:
step 1, data preparation: collecting original CT data, separating a labeled data set and a non-labeled data set, and sketching the labeled data set;
step 2, feature extraction: constructing a pre-training network based on self-supervised learning according to the characteristics of CT data, inputting a label-free data set into the pre-training network for iterative training, and selecting an optimal pre-training model;
step 3, generating a segmentation model: and constructing a segmentation network according to the segmentation task, loading the trained self-supervision pre-training model into the segmentation network, inputting the labeled data set into the segmentation network for iterative training to select an optimal model, and finally testing and evaluating the segmentation performance of the model.
In the step 1, different from the traditional image segmentation method based on the deep neural network, the method performs model training by using a method of combining a pre-learning task and a segmentation task based on self-supervision learning, and the pre-learning task performs feature extraction by using unlabeled original data; in the segmentation task, based on the extracted features, the labeled data is used for training to generate a segmentation model. For the two tasks, the step firstly needs to prepare the data to be processed, and the data preparation specifically comprises the following steps:
step 1a, data acquisition: a physicist scans a target region of a patient by using a CT machine to obtain CT image data, and a plurality of continuous CT images are generated by one-time scanning;
step 1b, data division: dividing the obtained original CT data into a non-labeled data set and a labeled data set according to the ratio of 6:4, and respectively corresponding to a pre-learning task and a segmentation task;
step 1c, data processing: and (4) performing target region delineation on the labeled data set in the step 1b by a physicist according to pathological features of the case, and generating a data set containing target region segmentation labels.
Example 2
The embodiment is further optimized on the basis of embodiment 1, and specifically comprises the following steps:
in the step 2, the feature extraction mainly generates a pre-training model through an executed pre-learning task, and the pre-learning task mainly uses the self-carried features of the data as labels to design and train the network. The CT images generated in the same scan are continuous, and therefore there is a similarity in features between any two CT images in the same scan. Considering that the CT scanner generates a plurality of information labels when generating the CT images, including coordinate information of the CT images, which indicates the scanning positions of the CT images, as shown in fig. 2, it can be known that two CT images have a "relative distance" therebetween according to the coordinate Z pointing to the head. Thus, the deep nerves can extract image features of the CT by learning such "relative distance" between CT images. The method specifically comprises the following steps:
step 2a, constructing a pre-learning network: the pre-learning task inputs two different CT images belonging to a case each time, and the output result is a single numerical value each time, so that the pre-learning network is realized by combining two twin deep neural networks in a weight sharing mode, and the pre-learning network comprises two modules of an encoder and a fusion device, as shown in figure 3.
Step 2b, sample data set construction: in the step, the acquired CT data is required to be constructed into an available sample data set, the CT data acquired in the step 1 comprises a plurality of cases, N cases are assumed in the CT data to be processed, each case comprises M CT images, and the size of the constructed sample set is required to be equal to that of the CT data to be processed
Figure 150694DEST_PATH_IMAGE001
Each sample can be represented as a triplet slice1, slice2, Dis, where slice1 and slice2 represent two CT images of the same case, respectively, and Dis represents the relative distance between the two CT images, which is calculated as:
Figure 861161DEST_PATH_IMAGE002
wherein the content of the first and second substances,
Figure 870706DEST_PATH_IMAGE003
are the coordinate values of the CT image,
Figure 858384DEST_PATH_IMAGE004
is the relative distance between the two CT images;
step 2c, CT data preprocessing, namely performing data augmentation operation on the unlabeled CT image data obtained in the step 1, and realizing the data augmentation operation through random overturning, translation and rotation, so that the pre-learning model can learn the characteristics with stronger robustness;
step 2d, training a pre-learning model: and (4) the augmented CT data is calculated according to the following weight ratio of 8: and 2, dividing the proportion of the data into a training set and a test set, inputting the data of the training set into the pre-learning network designed in the step 2a, sequentially performing previous calculation, back propagation and repetition for multiple times until the output loss value is converged, and stopping training when the output on the test set is the minimum value in multiple times of training, thereby completing feature extraction and obtaining the pre-learning model.
In step 2a, the encoder mainly comprises a plurality of convolution layers and pooling layers, and respectively extracts the characteristics of two input CT images to obtain two decoded characteristic vectors;
the fusion device aims to calculate and fuse the feature vectors extracted by the two networks, the method adopted by the fusion is a vector maximum interpolation method, namely, the maximum value on each channel in the two vectors is selected as a fusion vector, and the calculation formula is as follows:
Figure 662392DEST_PATH_IMAGE005
wherein
Figure 984789DEST_PATH_IMAGE006
Represents the output vector after the two networks are merged,
Figure 532445DEST_PATH_IMAGE007
Figure 374630DEST_PATH_IMAGE008
respectively representing the output vectors of the twin deep neural network 1 and the twin deep neural network 2.
Example 3
The embodiment is further optimized on the basis of embodiment 1 or 2, and specifically comprises the following steps:
in step 3, a pre-training model is obtained after feature extraction is completed, the pre-training model carries shallow image features of CT data, and a small amount of data containing segmentation labels are input into a segmentation network for training based on the pre-training model to obtain a segmentation model, wherein the method specifically comprises the following steps:
step 3a, constructing a segmentation network, wherein the segmentation model network is designed based on a deep neural network and comprises an encoder, a random multi-scale module and a decoder; the structure of the encoder part of the split model network is consistent with that of the encoder of the pre-learning network, and the main structure is shown in fig. 4 as follows:
step 3b, loading a pre-learning model: because the structure of the encoder of the pre-learning network is consistent with that of the encoder of the segmentation network during network design, the pre-training model obtained in the step 2 is directly loaded into the segmentation network, so that the segmentation network can directly obtain the characteristics of the CT image and can be suitable for scenes with too little label data;
step 3c, CT data preprocessing: considering that the labeled CT data volume is small, in order to make the segmentation model effect more feasible, the labeled CT data set obtained in step 1 needs to be subjected to augmentation operation, which is mainly implemented by random inversion, translation and rotation;
step 3d, training a segmentation model: the method comprises the following steps of performing model training by using a deep neural network for image segmentation; firstly, the data after the augmentation operation is divided into 7: 1: 2, dividing the proportion of the training set, the verification set and the test set, inputting the training set data into the network designed in the step 3a, sequentially performing previous item calculation, back propagation and repetition for multiple times until the loss value output by the network is converged, and stopping training when the output on the verification set is the minimum value in multiple times of training, thereby obtaining the optimal model result of the segmentation model network;
step 3e, segmentation model testing: after the training of the segmentation network model is completed, the segmentation effect of the segmentation model on the test set in (3d) needs to be quantitatively evaluated, and commonly used evaluation indexes include TPVF, PPV, and DSC, which are defined as follows:
Figure 615119DEST_PATH_IMAGE009
Figure 159233DEST_PATH_IMAGE010
Figure 776159DEST_PATH_IMAGE011
wherein
Figure 469921DEST_PATH_IMAGE012
And
Figure 146890DEST_PATH_IMAGE013
the model is used for evaluating the segmentation result, and the model is used for evaluating the segmentation result and is mainly used for evaluating the TPVF and the PPV.
In step 3a, the encoder is consistent with the encoder of the pre-training network in structure, and the segmentation network encoder is composed of a series of convolution and pooling layers and is used for extracting abstract features from the input CT image.
In step 3a, the random multi-scale module mainly comprises a cavity convolution module with 4 different cavity rates and a global average pooling layer, and the random multi-scale module further processes the feature vector output by the encoder to enable the feature vector to have features with different scales, so that the robustness of the model to the features with different scales is improved.
In step 3a, the decoder mainly aims to restore the features of the previous step to the original input size, generate the class prediction on each pixel of the original picture, and the decoder integrates the deep-layer features and the shallow-layer features, so that the feature multiplexing of the model can be improved, and the convergence of the model can be accelerated.

Claims (6)

1. A radiotherapy target area automatic segmentation method based on self-supervision learning is characterized by comprising the following steps:
step 1, data preparation: collecting original CT data, separating a labeled data set and a non-labeled data set, and sketching the labeled data set;
step 2, feature extraction: constructing a pre-training network based on self-supervision learning, inputting a label-free data set into the pre-training network for iterative training, and selecting an optimal pre-training model; in step 2, the feature extraction mainly generates a pre-training model through an executed pre-learning task, and specifically comprises the following steps:
step 2a, constructing a pre-learning network: the pre-learning network is realized by combining two twin deep neural networks in a weight sharing mode, and comprises two modules, namely an encoder and a fusion device; inputting two different CT images belonging to a case into a pre-learning network every time, and outputting a result as a single numerical value every time;
step 2b, sample data set construction: the CT data acquired in the step 1 comprises a plurality of cases, N cases are assumed to exist in the CT data to be processed, each case comprises M CT images, and the size of the constructed sample set is equal to
Figure FDA0003101295490000011
Each sample can be represented as a triplet slice1, slice2, Dis, where slice1 and slice2 represent two CT images of the same case, respectively, and Dis represents the relative distance between the two CT images, which is calculated as:
Dis=|Zslice1-Zsliice2|
wherein Z is the coordinate value of the CT image, and Dis is the relative distance between the two CT images;
step 2c, CT data preprocessing, namely performing data augmentation operation on the label-free CT image data obtained in the step 1 in a random overturning, translation and rotation mode;
step 2d, pre-learning model training: and (4) the augmented CT data is calculated according to the following weight ratio of 8: 2, inputting the training set data into the pre-learning network designed in the step 2a, sequentially performing previous item calculation, back propagation and repetition for multiple times until the output loss value is converged, and stopping training when the output on the test set is the minimum value in multiple times of training, thereby obtaining a pre-learning model by completing feature extraction;
step 3, generating a segmentation model: constructing a segmentation network, loading the trained self-supervision pre-training model into the segmentation network, and inputting the labeled data set into the segmentation network for iterative training to select an optimal model;
in step 3, a pre-training model is obtained after feature extraction is completed, the pre-training model carries shallow image features of CT data, and a small amount of data containing segmentation labels are input into a segmentation network for training based on the pre-training model to obtain a segmentation model, wherein the method specifically comprises the following steps:
step 3a, constructing a segmentation network, wherein the segmentation model network is designed based on a deep neural network and comprises an encoder, a random multi-scale module and a decoder;
step 3b, loading a pre-learning model: when the network is designed, the structures of the encoder of the pre-learning network and the encoder of the segmentation network are consistent, so that the pre-training model obtained in the step 2 is directly loaded into the segmentation network, and the segmentation network directly obtains the CT image characteristics;
step 3c, carrying out data augmentation operation on the label-free CT image data obtained in the step 1 in a random overturning, translation and rotation mode through CT data preprocessing;
step 3d, training a segmentation model: the method comprises the following steps of performing model training by using a deep neural network for image segmentation; firstly, the data after the augmentation operation is divided into 7: 1: 2, dividing the proportion of the training set, the verification set and the test set, inputting the training set data into the network designed in the step 3a, sequentially performing previous item calculation, back propagation and repetition for multiple times until the loss value output by the network is converged, and stopping training when the output on the verification set is the minimum value in multiple times of training, thereby obtaining the optimal model result of the segmentation model network;
step 3e, segmentation model testing: after the training of the segmentation network model is completed, the segmentation effect of the segmentation model on the test set in step 3d needs to be quantitatively evaluated, and common evaluation indexes include TPVF, PPV, and DSC, which are defined as follows:
Figure FDA0003101295490000021
Figure FDA0003101295490000022
Figure FDA0003101295490000023
wherein VsAnd VGRespectively representing the number of positive sample pixels predicted by the model and the number of real positive sample pixels, wherein TPVF represents the number of correct predictions in all real positive sample pixels, PPV represents the number of real positive samples in all predicted positive sample pixels, and DSC is a common index for evaluating segmentation results, and the common index mainly balances indexes of TPVF and PPV;
and 4, inputting the acquired CT data into a segmentation model, and outputting the automatically segmented CT data.
2. The method of claim 1, wherein the step 1, the data preparation comprises the following steps:
step 1a, data acquisition: CT image data are obtained by scanning the target region of a patient by using a CT machine, and a plurality of continuous CT images are generated by one-time scanning;
step 1b, data division: dividing the obtained original CT data into a non-labeled data set and a labeled data set according to the ratio of 6:4, and respectively corresponding to a pre-learning task and a segmentation task;
step 1c, data processing: and (4) according to pathological features of the case, performing target region delineation on the labeled data set in the step 1b, and generating a data set containing target region segmentation labels.
3. The automatic segmentation method for radiotherapy target area based on self-supervised learning of claim 1, wherein in step 2a, the encoder is composed of a plurality of convolution layers and pooling layers, and respectively extracts the features of two input CT images to obtain two decoded feature vectors;
the fusion device aims to calculate and fuse the feature vectors extracted by the two networks, the method adopted by the fusion is a vector maximum interpolation method, namely, the maximum value on each channel in the two vectors is selected as a fusion vector, and the calculation formula is as follows:
βcombine=Combine(Max(β1),Max(β2))
wherein beta iscombineRepresenting the output vector, beta, after the fusion of two networks1、β2Respectively representing the output vectors of the twin deep neural network 1 and the twin deep neural network 2.
4. The method of claim 1, wherein in step 3a, the encoder is consistent with the structure of the encoder of the pre-training network, and the encoder of the segmentation network is composed of a series of convolution and pooling layers for extracting abstract features from the input CT image.
5. The method of claim 1, wherein in step 3a, the stochastic multi-scale module comprises a cavity convolution module with 4 different cavity rates and a global average pooling layer, and the feature vectors output by the encoder are further processed.
6. The method of claim 1, wherein in step 3a, the decoder restores the previous feature to the original size, generates a class prediction for the original picture at each pixel, and fuses the deep features and the shallow features.
CN202110274005.3A 2021-03-15 2021-03-15 Automatic radiotherapy target area segmentation method based on self-supervision learning Active CN112686898B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110274005.3A CN112686898B (en) 2021-03-15 2021-03-15 Automatic radiotherapy target area segmentation method based on self-supervision learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110274005.3A CN112686898B (en) 2021-03-15 2021-03-15 Automatic radiotherapy target area segmentation method based on self-supervision learning

Publications (2)

Publication Number Publication Date
CN112686898A CN112686898A (en) 2021-04-20
CN112686898B true CN112686898B (en) 2021-08-13

Family

ID=75455566

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110274005.3A Active CN112686898B (en) 2021-03-15 2021-03-15 Automatic radiotherapy target area segmentation method based on self-supervision learning

Country Status (1)

Country Link
CN (1) CN112686898B (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113379764B (en) * 2021-06-02 2023-06-02 厦门理工学院 Pathological image segmentation method based on domain antagonism self-supervision learning
CN113435480B (en) * 2021-06-07 2022-06-21 电子科技大学 Method for improving long tail distribution visual recognition capability through channel sequential switching and self-supervision
CN114092446A (en) * 2021-11-23 2022-02-25 中国人民解放军总医院 Intracranial hemorrhage parameter acquisition method and device based on self-supervision learning and M-Net
CN114240966B (en) * 2021-12-13 2024-03-15 西北工业大学 Self-supervision learning method for 3D medical image segmentation training feature extractor
CN114298910B (en) * 2021-12-30 2024-01-23 上海商学院 MR-guided brachytherapy applicator segmentation method
CN115100390B (en) * 2022-08-24 2022-11-18 华东交通大学 Image emotion prediction method combining contrast learning and self-supervision region positioning

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108229479A (en) * 2017-08-01 2018-06-29 北京市商汤科技开发有限公司 The training method and device of semantic segmentation model, electronic equipment, storage medium
CN108256408A (en) * 2017-10-25 2018-07-06 四川大学 A kind of stem cell method for tracing based on deep learning
CN108985330A (en) * 2018-06-13 2018-12-11 华中科技大学 A kind of autoencoder network and its training method abnormal use electric detection method and system
CN109697469A (en) * 2018-12-26 2019-04-30 西北工业大学 A kind of self study small sample Classifying Method in Remote Sensing Image based on consistency constraint
CN110378044A (en) * 2019-07-23 2019-10-25 燕山大学 Multiple Time Scales convolutional neural networks flexible measurement method based on attention mechanism
CN110503053A (en) * 2019-08-27 2019-11-26 电子科技大学 Human motion recognition method based on cyclic convolution neural network
CN111046855A (en) * 2019-06-21 2020-04-21 腾讯科技(深圳)有限公司 Picture identification method, identification model training method and device and electronic device
CN111858933A (en) * 2020-07-10 2020-10-30 暨南大学 Character-based hierarchical text emotion analysis method and system
CN112164447A (en) * 2020-10-20 2021-01-01 腾讯科技(深圳)有限公司 Image processing method, device, equipment and storage medium
CN112270660A (en) * 2020-09-30 2021-01-26 四川大学 Nasopharyngeal carcinoma radiotherapy target area automatic segmentation method based on deep neural network
CN112419452A (en) * 2020-12-24 2021-02-26 福州大学 Rapid merging system and method for PD-L1 digital pathological section images of stomach cancer

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102842132B (en) * 2012-07-12 2015-12-09 上海联影医疗科技有限公司 A kind of CT pulmonary nodule detection method
WO2014049667A1 (en) * 2012-09-28 2014-04-03 株式会社島津製作所 Digital image processing method and imaging device
CN104616307B (en) * 2015-02-12 2018-05-08 河北大学 A kind of lung CT image adhesion vascular type nodule detection methods
CN108364294B (en) * 2018-02-05 2021-01-12 西北大学 Multi-organ segmentation method for abdominal CT image based on superpixels
CN110610472A (en) * 2018-06-16 2019-12-24 深圳市前海安测信息技术有限公司 Computer device and method for realizing classification detection of lung nodule images
CN109409503B (en) * 2018-09-27 2020-07-24 深圳市铱硙医疗科技有限公司 Neural network training method, image conversion method, device, equipment and medium
CN110458852B (en) * 2019-08-13 2022-10-21 四川大学 Lung tissue segmentation method, device and equipment based on capsule network and storage medium
CN111104984B (en) * 2019-12-23 2023-07-25 东软集团股份有限公司 Method, device and equipment for classifying CT (computed tomography) images
CN111738997A (en) * 2020-06-11 2020-10-02 浙江大学 Method for calculating new coronary pneumonia lesion area ratio based on deep learning

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108229479A (en) * 2017-08-01 2018-06-29 北京市商汤科技开发有限公司 The training method and device of semantic segmentation model, electronic equipment, storage medium
CN108256408A (en) * 2017-10-25 2018-07-06 四川大学 A kind of stem cell method for tracing based on deep learning
CN108985330A (en) * 2018-06-13 2018-12-11 华中科技大学 A kind of autoencoder network and its training method abnormal use electric detection method and system
CN109697469A (en) * 2018-12-26 2019-04-30 西北工业大学 A kind of self study small sample Classifying Method in Remote Sensing Image based on consistency constraint
CN111046855A (en) * 2019-06-21 2020-04-21 腾讯科技(深圳)有限公司 Picture identification method, identification model training method and device and electronic device
CN110378044A (en) * 2019-07-23 2019-10-25 燕山大学 Multiple Time Scales convolutional neural networks flexible measurement method based on attention mechanism
CN110503053A (en) * 2019-08-27 2019-11-26 电子科技大学 Human motion recognition method based on cyclic convolution neural network
CN111858933A (en) * 2020-07-10 2020-10-30 暨南大学 Character-based hierarchical text emotion analysis method and system
CN112270660A (en) * 2020-09-30 2021-01-26 四川大学 Nasopharyngeal carcinoma radiotherapy target area automatic segmentation method based on deep neural network
CN112164447A (en) * 2020-10-20 2021-01-01 腾讯科技(深圳)有限公司 Image processing method, device, equipment and storage medium
CN112419452A (en) * 2020-12-24 2021-02-26 福州大学 Rapid merging system and method for PD-L1 digital pathological section images of stomach cancer

Non-Patent Citations (8)

* Cited by examiner, † Cited by third party
Title
"Improving Cytoarchitectonic Segmentation of Human Brain Areas with Self-suprevised Siamese Networks;Hannah Spitzer等;《arXiv:1806.05104v1》;20180613;正文第1-4节,图1-图2、图4 *
Automated Segmentation of the Clinical Target Volume in the Planning CT for Breast Cancer Using Deep Neural Networks;Xiaofeng Qi等;《IEEE TRANSACTIONS ON CYBERNETICS》;20200830;1-11 *
IFSS-Net: Interactive Few-Shot Siamese etwork for Faster Muscle Segmentation and Propagation in Volumetric Ultrasound;Dawood Al Chanti等;《IEEE TRANSACTIONS ON MEDICAL IMAGING》;20210209;1-14 *
Rubik’s Cube+: A self-supervised feature learning framework for 3D medical image analysis;Jiuwen Zhu等;《Medical Image Analysis》;20200606;1-11 *
SELF-SUPERVISED LEARNING FOR SEGMENTATION;bhinav Dhere等;《arXiv:2101.05456v1》;20210114;1-4 *
基于自监督学习的卷积神经网络在CT图像中的肝脏自动分割研究;刘祥强等;《科学技术创新》;20210122(第03期);78-79 *
基于自监督学习的病理图像层次分割;吴崇数等;《计算机应用》;20200610;第40卷(第6期);1856-1862 *
面向乳腺超声图像分割的混合监督双通道反馈 U-Net;贡荣麟等;《中国图象图形学报》;20201016;第25卷(第10期);2206-2217 *

Also Published As

Publication number Publication date
CN112686898A (en) 2021-04-20

Similar Documents

Publication Publication Date Title
CN112686898B (en) Automatic radiotherapy target area segmentation method based on self-supervision learning
CN113077471B (en) Medical image segmentation method based on U-shaped network
Bielski et al. Emergence of object segmentation in perturbed generative models
CN110458844B (en) Semantic segmentation method for low-illumination scene
CN111612807B (en) Small target image segmentation method based on scale and edge information
CN110930421A (en) Segmentation method for CBCT (Cone Beam computed tomography) tooth image
CN110992351B (en) sMRI image classification method and device based on multi-input convolution neural network
CN113012172A (en) AS-UNet-based medical image segmentation method and system
CN113362271B (en) Blood vessel three-dimensional image segmentation method and device, computer equipment and storage medium
Shu et al. LVC-Net: Medical image segmentation with noisy label based on local visual cues
Wazir et al. HistoSeg: Quick attention with multi-loss function for multi-structure segmentation in digital histology images
CN111652231B (en) Casting defect semantic segmentation method based on feature self-adaptive selection
CN114119525A (en) Method and system for segmenting cell medical image
CN115293986A (en) Multi-temporal remote sensing image cloud region reconstruction method
CN116012255A (en) Low-light image enhancement method for generating countermeasure network based on cyclic consistency
CN112329662B (en) Multi-view saliency estimation method based on unsupervised learning
CN117557683A (en) Novel image generation method driven by text and semantic segmentation map together
CN116912727A (en) Video human behavior recognition method based on space-time characteristic enhancement network
Kulikajevas et al. Adversarial 3D human pointcloud completion from limited angle depth data
Wicaksono et al. Improve image segmentation based on closed form matting using K-means clustering
Zhou et al. Cross-modal guidance for hyperfluorescence segmentation in fundus fluorescein angiography
CN111428734B (en) Image feature extraction method and device based on residual countermeasure inference learning and computer readable storage medium
Hossam et al. A Review of Generative Adversarial Networks Applications
CN113487622B (en) Head-neck organ image segmentation method, device, electronic equipment and storage medium
Ning et al. Doppler Image-Based Weakly-Supervised Vascular Ultrasound Segmentation with Transformer

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant