CN108806792B - Deep learning face diagnosis system - Google Patents

Deep learning face diagnosis system Download PDF

Info

Publication number
CN108806792B
CN108806792B CN201711255031.1A CN201711255031A CN108806792B CN 108806792 B CN108806792 B CN 108806792B CN 201711255031 A CN201711255031 A CN 201711255031A CN 108806792 B CN108806792 B CN 108806792B
Authority
CN
China
Prior art keywords
neural network
convolutional neural
module
deep learning
layer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201711255031.1A
Other languages
Chinese (zh)
Other versions
CN108806792A (en
Inventor
金波
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Publication of CN108806792A publication Critical patent/CN108806792A/en
Application granted granted Critical
Publication of CN108806792B publication Critical patent/CN108806792B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/243Classification techniques relating to the number of classes
    • G06F18/24323Tree-organised classifiers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention provides a facial diagnosis method and system based on deep learning. The invention uses a large number of face images with disease labels to train the deep learning network, and after repeated iteration and parameter updating, the deep learning network can automatically and effectively extract the face characteristics including eyes, ears, noses, mouths and eyebrows, and find out the internal relation between the face characteristics and various diseases of human bodies, thereby effectively detecting and screening the diseases of new face images of testees, and the invention is a non-invasive computer-aided automatic diagnosis method. The system can update parameters by continuously collecting the face pictures of the patient and then training, so that the prediction judgment result is more accurate. The method and the system can effectively solve the problem of difficult disease detection in poverty-stricken laggard areas, so that people can conveniently and quickly carry out non-invasive automatic disease detection and screening through terminals such as mobile phones, computers and the like, and diseases can be treated in time, thereby improving the living quality level of human beings.

Description

Deep learning face diagnosis system
Technical Field
The invention belongs to the fields of computer vision, machine learning and medicine, and particularly relates to a computer method and a system for diagnosing diseases through facial diagnosis based on deep learning.
Background
Over two thousand years ago, the Chinese ancient book Huangdi's internal classic records that the blood and qi flow from the twelve main meridians and the three hundred and sixty five main meridians to the face and the orifices. This indicates that pathological changes of the five zang-organs and six fu-organs of a person are manifested in the relevant areas of the face. In China, an experienced doctor can grasp the whole body and local lesion conditions of a patient by observing facial features of the patient, and the diagnosis mode is called 'facial diagnosis'. The disadvantage of the diagnostic method is that it requires a lot of experience of the doctor to have a high accuracy.
With the development of science and technology, big data and deep learning technology develop rapidly in recent years. The deep learning technology is that a multi-level neural network is utilized, and a large amount of data are trained, so that a computer can learn and understand complex data such as images and sounds and can make corresponding behaviors, which is sometimes better than that of a human being. The deep learning network can extract contrast features which are high in complexity and difficult for human beings to understand.
In 2006, Geoffrey Hinton, professor Geoffrey, canada, proposed an algorithm for deep learning. He and his student Ruslan Salakhutdinov published a paper in science, from which he triggered a hot tide of deep learning studies. In 2012, the deep convolutional neural network structure AlexNet proposed by the university of toronto team won the champion in the ImageNet large-scale visual recognition challenge race (ILSVRC). In 2014, Google corporation and oxford university visual geometry groups each achieved excellent performance in ILSVRC using their deep convolutional neural networks GoogleNet and VGGNet, which are more structurally complex than AlexNet. In 2015, Revzmin et al, Microsoft Asian institute, using the residual error network ResNet, captured the ILSVRC champion, and obtained the best results of the year in image classification, target detection, semantic segmentation, and other tasks. The residual network has a greater depth, up to 152 layers. In 2016, the integrated deep neural network performed well in ILSVRC tasks.
Therefore, in the current big data era, people can utilize increasingly complex depth models to analyze and extract effective information contained in mass data, so that more accurate prediction can be made on unknown data. The essence of deep learning is that learning features are extracted by constructing a machine learning model with a plurality of hidden layers and a large amount of training data, so that the accuracy of classification or prediction is improved. A large number of research experiments show that the characteristic learning method based on big data deep learning can be used for more effectively expressing the intrinsic information of the data than the method of artificially constructing the characteristic. The invention is a method for realizing the purpose of diagnosing diseases through the face by utilizing big data deep learning.
Disclosure of Invention
The invention aims to conveniently and rapidly detect and screen diseases for all human beings and improve the quality level of life of the human beings, in particular to poor and lagged regions.
In order to achieve the purpose, the inventor provides a method and a system for non-invasive automatic detection and screening of diseases only through face images based on a big data deep learning technology. The method comprises the following steps:
s100: and collecting a proper face image sample with a disease diagnosis result label, and establishing a face image database with the disease diagnosis label.
S200: and combining a convolution layer, a pooling layer, a batch normalization function ReLu, a residual error structure and an inclusion structure into a mixing module, then combining a dimensionality reduction module, and establishing a deep convolutional neural network by a plurality of combinations in a cascade mode.
S300: and (5) preprocessing the picture acquired in the S100, taking the preprocessed picture as training data, and training a deep convolutional neural network model to obtain network model parameters.
S400: and (5) extracting the features of the training data by using the trained deep convolutional neural network model in the step (S300), and training a random forest classifier by using the extracted features to obtain a classification model.
S500: inputting the attempted image, extracting the features of the human face image to be classified by using the convolutional neural network model in S300, and then obtaining the disease category to which the human face image to be detected belongs by using the classifier model obtained in S400.
The invention is also characterized in that:
step S100 is to add a disease diagnosis result label according to a disease diagnosis report issued at that time in a hospital while acquiring a face image of a subject. And establishing a database of Niger species (black species), Mongolian species (yellow species) and Caucasian species (caucasian species) according to different skin colors. Preprocessing the collected pictures, wherein the preprocessed pictures have the same size and are input in three color dimensions of red, green and blue.
Step S200 uses a hybrid module combining a residual structure and an inclusion structure in the network construction. The input data is processed by Batch Normalization (Batch Normalization) in the network construction, and the pooling layer uses a global average pooling method.
In step S300, the preprocessed face image is divided into two parts, namely a training sample and a testing sample, the deep convolutional neural network is trained by the training sample, and the weight is updated by adopting a supervised training mode and a back propagation algorithm. And training the deep convolutional neural network using a Graphics Processor (GPU). And finally, inputting the test sample into a deep convolution neural network for testing. If the pre-trained network is adopted, the fine-tuning technology is required to be assisted.
In step S400, the trained deep convolutional neural network model is used as a feature extractor, and a classifier is trained by using the extracted features to obtain a classification model. This step of the function requires the use of a multivariate classifier model.
The disease types which can be detected and screened in step S500 include hyperthyroidism, hypothyroidism, hansheng disease, acromegaly, renal failure, neonatal pathological jaundice, Down 'S syndrome, thalassemia, DiGeorge' S syndrome and mitral stenosis. More and more disease species will be added.
The invention has the beneficial effect that the invention can automatically extract the characteristics under the data driving of the training image with the disease diagnosis result label through the deep convolutional neural network. The deep convolutional neural network can avoid the additional preprocessing process of the traditional image recognition algorithm. A residual structure is applied in this deep convolutional neural network to solve the problem of accuracy degradation as the network deepens.
The features extracted by the deep convolutional neural network can be prevented from being influenced by color illumination and the like, and more complex features which are difficult to describe by human can be extracted. The deep convolutional neural network and the classifiers are effectively combined, so that not only can rich characteristics be extracted, but also a proper classifier can be selected according to the characteristics of a data set of the classifier, and the accuracy of the whole classification model is improved.
Drawings
FIG. 1 is a schematic diagram of the overall machine learning algorithm of the present invention;
FIG. 2 is a schematic diagram of the structure of the deep convolutional neural network of the present invention;
fig. 3 is a schematic diagram of the system architecture of the present invention.
Detailed Description
In order to explain technical contents, structural features, and objects and effects of the present invention in detail, the following detailed description is given with reference to the accompanying drawings in conjunction with the embodiments.
Referring to fig. 1, in order to solve the technical problem of the present invention, one technical solution adopted by the present invention is to provide a method based on deep learning, including the following steps:
s100: collecting a proper face image sample, adding a disease type label according to a disease diagnosis report issued by a hospital during collection, and establishing a face image database with a disease diagnosis result label. Databases are divided into three major categories, the Niger species (black), Mongolian species (yellow), and Caucasian species (caucasian). Preprocessing the collected pictures, wherein the preprocessed pictures have red, green and blue three-color dimension input. The matconvnet code based on the Matlab platform below is an example of a preprocessing step.
The read image is first normalized and processed,
img0=imresize(img0,net.meta.normalization.imageSize(1:2));
the mean value of the image is then subtracted,
img0=img0-net.meta.normalization.averagelmage;
s200: and (3) establishing a deep convolutional neural network by utilizing a convolutional layer, a pooling layer, a batch normalization and activation function ReLu assisted by a residual error structure and an inclusion structure.
The GPU has excellent parallel computing capability, so that great help can be provided for big data computation, and in the process, the GPU is used for training and deducing the deep convolutional neural network.
And (3) rolling layers: the input images are convoluted by using different convolution kernels to obtain a feature map, and each convolution kernel detects specific features on all positions on the input feature map, so that weight sharing on the same input feature map is realized. Each input feature map element is x, the weight connecting the input feature map element and the output feature map element is w, the output feature map element is y, the bias parameter is b, f is an activation function, and the formula is as follows:
Figure GSB0000196792270000041
the activation function adopts ReLu, and the formula is as follows:
f(x)=max(0,x)
a pooling layer: unimportant samples in the feature map can be removed through down-sampling, training parameters are reduced, and therefore the degree of fitting of the over-model is reduced. Here we use global average pooling (global average pooling) to average the ensemble for each feature map.
Random deactivation layer (Dropout): in the training process of the deep learning network, the neural unit of the deep learning network is temporarily discarded from the network according to a certain probability, so that overfitting can be effectively prevented, and the value is 0.8.
Residual structure: the residual structure means that each residual module has two paths, one path is an input feature, the other path is a residual obtained by performing convolution operation on the input feature for multiple times, and finally, the input feature and the residual are added:
H(x)=F(x)+x
x represents input features, f (x) represents residual features, and h (x) represents residual module output features.
The inclusion structure: the structure formed by convolution layers (such as 1 × 1, 3 × 3) with different sizes in parallel increases the width of the neural network and the adaptability of the network nerves to the scale.
The Hybrid module combining the residual structure and the inclusion structure makes the neural network deeper and wider.
S300: and (5) taking the picture learning sample obtained in the step (S100) as training data, training a convolutional neural network model, and obtaining network model parameters. In the process of training the convolutional neural network, a supervised training mode and a back propagation algorithm are adopted to update the weight. The process is as follows:
calculating the error between the sample output characteristic and the theoretical label value:
Figure GSB0000196792270000051
the error is transmitted to each layer along the reverse path of the convolutional neural network, and the weight is updated according to a gradient descent algorithm:
Figure GSB0000196792270000052
eta is the learning rate, delta is the error term, and the weight update formula is as follows:
Δwij=ηδjxij
with the continuous update of the convolution kernel weight value, the output error of the convolution neural network is continuously reduced and finally tends to be stable. If the pre-trained network is adopted, the fine-tuning technology is required to be assisted.
S400: and (5) extracting the characteristics of the training data by using the trained convolutional neural network model in the step (S300), and training by using a multivariate classifier to obtain a multivariate classification model. Here, a random forest algorithm is used as an example, and a training set is used to generate a plurality of classification decision trees, and the decision trees form a random forest. In the prediction process, each tree can predict a result, and finally, the class with the most classified results in all the decision trees is taken as the final result, the splitting index adopted for generating the classified decision trees is a kini coefficient, and the definition formula is as follows:
Figure GSB0000196792270000061
in the classification problem, there is a class k, and the probability that a sample point belongs to the kth class is pkThe sample capacity is | D |, | C |kIs of class CkThe number of samples.
S500: referring to fig. 3, a human face image is input by a subject at a terminal, the human face image is transmitted to a cloud computing platform, features of the human face image to be classified are extracted by using a convolutional neural network model in S300, then a classifier model obtained in S400 is used to obtain a disease class to which the human face image to be classified belongs, and the disease class is returned to the subject. The disease types which can be detected and screened in S500 at present comprise hyperthyroidism, hypothyroidism, hansheng disease, acromegaly, renal failure, neonatal pathological jaundice, Down' S syndrome, thalassemia, DiGeorge syndrome and mitral stenosis. More and more disease categories will be added to this invention in the future.
The above description is only an embodiment of the present invention, and not intended to limit the scope of the present invention, and all modifications of equivalent structures and equivalent processes performed by the present specification and drawings, or directly or indirectly applied to other related technical fields, are included in the scope of the present invention.

Claims (5)

1. A deep learning face diagnosis system, which detects and screens diseases through human face features, is characterized by comprising the following steps:
s100: collecting a face image sample with a disease diagnosis result label, and establishing a face image database with the disease diagnosis label, wherein the types of the disease diagnosis label comprise hyperthyroidism, hypothyroidism, hansheng disease, acromegaly, renal failure, neonatal pathological jaundice, down syndrome, thalassemia, diqiao syndrome and mitral stenosis;
s200: constructing a deep convolutional neural network, wherein the neural network comprises an input layer, a convolutional layer, a first pooling layer, a plurality of combination modules, a second pooling layer, a random inactivation layer (dropout) and a classification layer which are sequentially connected; the combined module consists of a mixing module followed by a dimension reduction (reduction) module; in the mixing module, an excitation function layer is connected firstly, then the mixing module is divided into two paths, one path is a direct output characteristic, the other path is divided into three parallel sub-paths, each sub-path comprises convolution layers with different numbers, the outputs of the three parallel sub-paths are spliced together to be used as the input of the other convolution layer, finally the direct output characteristic and the output characteristic of the other convolution layer are added, and then an excitation function is used for processing; the mixing module can increase the adaptability to the scale, avoid the accuracy reduction along with the deepening of the network and increase the depth and the width of the network structure;
s300: preprocessing the image acquired in the step S100, taking the preprocessed image and a label corresponding to the preprocessed image as training data, and training a deep convolutional neural network model to obtain network model parameters; extracting the characteristics of training data by using a trained S200 deep convolutional neural network model, and carrying out supervised training on a random forest classifier by using the extracted characteristics and a corresponding label to obtain a random forest classification model;
s400: and inputting a face image by a subject at a terminal, transmitting the face image to a cloud computing platform, extracting the features of the face image to be classified by using the trained deep convolutional neural network, obtaining the disease category to which the face image to be classified belongs by using the forest classifier model obtained in the S300, and returning the disease category to the subject.
2. The deep learning interview system of claim 1 wherein said step S100 further comprises categorizing the database into three broad categories according to skin tone: a database of Niger Roman race, Mongolian race, and Caucasian race; the preprocessed images in step S300 all have the same size and three color dimension inputs of red, green and blue,
the preprocessing step of matconvnet codes based on the Matlab platform comprises the following steps:
firstly, normalizing the read image:
img0=imresize(img0,net.meta.normalization.imageSize(1∶2))
the image mean is then subtracted:
img0=img0-net.meta.normalization.averageImage。
3. the deep learning face diagnosis system according to claim 1, wherein the implementation of step S200 uses a hybrid module combining a residual structure and an inclusion structure in the network construction, and the output formula of the residual structure is as follows:
H(x)=F(x)+x
x represents input features, F (x) represents residual features, H (x) represents residual module output features;
the output formula of the convolution layer in the deep convolutional neural network is as follows:
Figure FSB0000196792260000021
each input characteristic diagram element is x, the weight value connecting the input characteristic diagram element and the output characteristic diagram element is w, the output characteristic diagram element is y, the bias parameter is b, and f is an activation function;
the ReLu function is used as the activation function in the deep convolutional neural network, and the output formula is as follows:
f(x)=max(0,x)
the second pooling layer uses a global average pooling method, which is to calculate an overall average value for each feature map, and the output formula is as follows:
Figure FSB0000196792260000022
wherein, ykRepresenting the global average pooled output value, x, from the kth feature mapkpqRepresents the element at (p, q) in the kth feature map region R, | R | represents the number of all elements of the kth feature map;
the purpose of the random deactivation layer (dropout) is to prevent overfitting.
4. The deep learning surface diagnosis system of claim 1, wherein the deep learning surface diagnosis system of step S300 is implemented by using a Graphics Processing Unit (GPU) to train a deep convolutional neural network by adopting a supervised training mode and a back propagation algorithm to update weights, and the process is as follows:
calculating the error between the sample output characteristic and the theoretical label value:
Figure FSB0000196792260000031
the error is transmitted to each layer along the reverse path of the convolutional neural network, and the weight is updated according to a gradient descent algorithm:
Figure FSB0000196792260000032
eta is the learning rate, delta is the error term, and the weight update formula is as follows:
Δwij=ηδjxij
extracting image features and label training by using a convolutional neural network to generate a plurality of classification decision trees, wherein the decision trees form a random forest; during prediction, each tree can predict a result, and finally, the class with the most classified results in all the decision trees is taken as a final result, and the splitting index adopted for generating the classified decision trees is a kini coefficient:
Figure FSB0000196792260000033
the probability that a sample point belongs to class k is pkThe sample capacity is | D |, | C |kIs of class CkIn this embodiment, k is 11.
5. The deep learning face diagnosis system according to claim 1, wherein the system comprises a terminal and a cloud computing platform; the terminal comprises a face image acquisition module, an information sending module and an information receiving module; the cloud computing platform comprises an image receiving module, a face image database with a disease diagnosis result label, a deep convolutional neural network training module, a deep convolutional neural network classification prediction module and a classification prediction result information sending module; the user can input the face image to the cloud computing platform through the terminal equipment including a desktop computer, a smart phone, a tablet computer and a notebook computer, the facial features are automatically extracted through the deep convolutional neural network, the classification model is used for disease detection, and the system returns a disease detection classification result to the terminal equipment.
CN201711255031.1A 2017-05-03 2017-11-25 Deep learning face diagnosis system Active CN108806792B (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN2017103453755 2017-05-03
CN201710345375 2017-05-03

Publications (2)

Publication Number Publication Date
CN108806792A CN108806792A (en) 2018-11-13
CN108806792B true CN108806792B (en) 2022-01-04

Family

ID=64095168

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711255031.1A Active CN108806792B (en) 2017-05-03 2017-11-25 Deep learning face diagnosis system

Country Status (1)

Country Link
CN (1) CN108806792B (en)

Families Citing this family (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109741823A (en) * 2018-11-23 2019-05-10 杭州电子科技大学 A kind of pneumothorax aided diagnosis method based on deep learning
JP7395604B2 (en) * 2018-12-04 2023-12-11 ジャン・ルオウェイ Automatic image-based skin diagnosis using deep learning
CN109685968A (en) * 2018-12-15 2019-04-26 西安建筑科技大学 A kind of the identification model building and recognition methods of the banknote image defect based on convolutional neural networks
CN109766920A (en) * 2018-12-18 2019-05-17 任飞翔 Article characteristics Model Calculating Method and device based on deep learning
CN109785311B (en) * 2019-01-14 2021-06-04 深圳和而泰数据资源与云技术有限公司 Disease diagnosis device, electronic equipment and storage medium
CN109920538B (en) * 2019-03-07 2022-11-25 中南大学 Zero sample learning method based on data enhancement
CN111951950B (en) * 2019-05-16 2023-11-24 中国科学院自动化研究所 Three-dimensional data medical classification system based on deep learning
CN110363075B (en) * 2019-06-03 2020-03-20 南京特殊教育师范学院 Suspicious ill face detection system based on big data server
CN111353520A (en) * 2019-07-03 2020-06-30 孙彤 Method for identifying vermilion ink time sequence
CN111104961A (en) * 2019-10-31 2020-05-05 太原理工大学 Method for classifying breast cancer based on improved MobileNet network
CN111489325B (en) * 2019-12-16 2023-07-28 杭州电子科技大学 Instant identification device for hyperthyroidism disease and control method thereof
CN111341459A (en) * 2020-02-28 2020-06-26 上海交通大学医学院附属上海儿童医学中心 Training method of classified deep neural network model and genetic disease detection method
CN111951934A (en) * 2020-08-20 2020-11-17 陈文立 Novel acromegaly screening system and screening method thereof
CN111803024B (en) * 2020-08-27 2022-10-04 张桦 System for measuring eyeball protrusion degree based on deep learning algorithm
CN112163512A (en) * 2020-09-25 2021-01-01 杨铠郗 Autism spectrum disorder face screening method based on machine learning
CN112687393A (en) * 2020-12-29 2021-04-20 康奥生物科技(天津)股份有限公司 Disease prediction system based on big data
CN113012823A (en) * 2021-03-30 2021-06-22 海口讯婕城医疗科技有限公司 Cloud platform system for electrocardiograph
CN113139597B (en) * 2021-04-19 2022-11-04 中国人民解放军91054部队 Statistical thought-based image distribution external detection method
CN113284613A (en) * 2021-05-24 2021-08-20 暨南大学 Face diagnosis system based on deep learning
CN113625094A (en) * 2021-09-11 2021-11-09 沈阳工业大学 Subway platform screen door fault diagnosis system
WO2023161956A1 (en) * 2022-02-28 2023-08-31 Saanvi Mehra Device and method for assessing a subject for down syndrome

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102426652A (en) * 2011-10-10 2012-04-25 北京工业大学 Traditional Chinese medicine face color identifying and retrieving method based on image analysis
CN102973253A (en) * 2012-10-31 2013-03-20 北京大学 Method and system for monitoring human physiological indexes by using visual information
JP2014182424A (en) * 2013-03-18 2014-09-29 Seiko Epson Corp Printing equipment, printing system, and printed matter
CN104699931A (en) * 2013-12-09 2015-06-10 广州华久信息科技有限公司 Neural network blood pressure prediction method and mobile phone based on human face
CN105590102A (en) * 2015-12-30 2016-05-18 中通服公众信息产业股份有限公司 Front car face identification method based on deep learning

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102426652A (en) * 2011-10-10 2012-04-25 北京工业大学 Traditional Chinese medicine face color identifying and retrieving method based on image analysis
CN102973253A (en) * 2012-10-31 2013-03-20 北京大学 Method and system for monitoring human physiological indexes by using visual information
JP2014182424A (en) * 2013-03-18 2014-09-29 Seiko Epson Corp Printing equipment, printing system, and printed matter
CN104699931A (en) * 2013-12-09 2015-06-10 广州华久信息科技有限公司 Neural network blood pressure prediction method and mobile phone based on human face
CN105590102A (en) * 2015-12-30 2016-05-18 中通服公众信息产业股份有限公司 Front car face identification method based on deep learning

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于面象特征的中医体质自动辨识***研究;梁玉梅;《中国优秀硕士学位论文全文数据库 医药卫生科技辑》;20170315(第3期);E056-67 *

Also Published As

Publication number Publication date
CN108806792A (en) 2018-11-13

Similar Documents

Publication Publication Date Title
CN108806792B (en) Deep learning face diagnosis system
Sahu et al. FINE_DENSEIGANET: Automatic medical image classification in chest CT scan using Hybrid Deep Learning Framework
Salido et al. Using deep learning to detect melanoma in dermoscopy images
CN110276745B (en) Pathological image detection algorithm based on generation countermeasure network
CN109117864A (en) Coronary heart disease risk prediction technique, model and system based on heterogeneous characteristic fusion
CN108257135A (en) The assistant diagnosis system of medical image features is understood based on deep learning method
CN109493342B (en) Skin disease picture lesion type classification method based on deep learning
CN111223553A (en) Two-stage deep migration learning traditional Chinese medicine tongue diagnosis model
CN107516312A (en) A kind of Chinese medicine complexion automatic classification method using shallow-layer neutral net
CN113610118B (en) Glaucoma diagnosis method, device, equipment and method based on multitasking course learning
Singh et al. Collaboration of features optimization techniques for the effective diagnosis of glaucoma in retinal fundus images
CN114398979A (en) Ultrasonic image thyroid nodule classification method based on feature decoupling
CN112668486A (en) Method, device and carrier for identifying facial expressions of pre-activated residual depth separable convolutional network
Elayaraja et al. An efficient approach for detection and classification of cancer regions in cervical images using optimization based CNN classification approach
CN111462082A (en) Focus picture recognition device, method and equipment and readable storage medium
CN117010971B (en) Intelligent health risk providing method and system based on portrait identification
CN110992309B (en) Fundus image segmentation method based on deep information transfer network
Cao et al. 3D convolutional neural networks fusion model for lung nodule detection onclinical CT scans
Kotiyal et al. Diabetic retinopathy binary image classification using PySpark
CN115661185A (en) Fundus image blood vessel segmentation method and system
Yasmin et al. Impact of fuzziness for skin lesion classification with transformer-based model
Wu et al. Mscan: Multi-scale channel attention for fundus retinal vessel segmentation
Charulatha et al. Computer-Aided Design for skin disease Identification and Categorization Using Image Processing
Hatode et al. Evolution and Testimony of Deep Learning Algorithm for Diabetic Retinopathy Detection
CN110517783A (en) Model training method, device and heart disease diagnosis device, equipment, medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant