CN111950736A - Migration ensemble learning method, terminal device, and computer-readable storage medium - Google Patents

Migration ensemble learning method, terminal device, and computer-readable storage medium Download PDF

Info

Publication number
CN111950736A
CN111950736A CN202010726052.2A CN202010726052A CN111950736A CN 111950736 A CN111950736 A CN 111950736A CN 202010726052 A CN202010726052 A CN 202010726052A CN 111950736 A CN111950736 A CN 111950736A
Authority
CN
China
Prior art keywords
migration
model
migratable
total
individual
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010726052.2A
Other languages
Chinese (zh)
Other versions
CN111950736B (en
Inventor
袁春
钟括
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen International Graduate School of Tsinghua University
Original Assignee
Shenzhen International Graduate School of Tsinghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen International Graduate School of Tsinghua University filed Critical Shenzhen International Graduate School of Tsinghua University
Priority to CN202010726052.2A priority Critical patent/CN111950736B/en
Publication of CN111950736A publication Critical patent/CN111950736A/en
Application granted granted Critical
Publication of CN111950736B publication Critical patent/CN111950736B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Medical Informatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Artificial Intelligence (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention provides a migration ensemble learning method, a terminal device and a computer readable storage medium, wherein the method comprises the following steps: s1: uniformly sampling at least 3 different total migratable degrees of the source models from 0 to 1, wherein each total migratable degree of the source models corresponds to one individual learner; s2: for each individual learner, determining an optimal migration strategy based on the total migratable extent of the corresponding source model; s3: migrating a source model to an individual learner of a target task sequentially through a parameter migration regular term and a task loss function based on the optimal migration strategy corresponding to each individual learner; s4: and obtaining an integrated model in an average mode based on the individual learners of each target task. By utilizing the effectiveness of ensemble learning in relieving overfitting, the estimation of the migration strategy is more stable and efficient through reasonable assumption, and finally the generalization capability of the ensemble model is greatly improved.

Description

Migration ensemble learning method, terminal device, and computer-readable storage medium
Technical Field
The invention relates to the technical field of artificial intelligence and transfer learning, in particular to a transfer ensemble learning method, a terminal device and a computer readable storage medium.
Background
The introduction of deep neural networks has significantly facilitated the development of computer vision and natural language processing in recent years. Compared with the traditional method for manually defining the characteristics, the model can automatically learn the characteristics with more discrimination through layer-by-layer stacking of the network. However, the neural network also brings a new problem that it requires a large amount of tag data. Therefore, to achieve better generalization performance, the model needs to be trained with more data. This seems to be a ferrology in the deep learning field today. However, obtaining a large amount of tag data is costly. For example, in the segmentation task, each pixel point needs to be labeled with a category, so that a large amount of manual labor is consumed. Even more fatal, due to privacy protection, it becomes completely impossible to collect a large amount of tag data.
Facing the problem of a small number of training samples, one current direction of fire comparison is transfer learning. Intuitively, the method of migrating and riding the bicycle can be used for learning how to ride the motorcycle and how to take the Chinese chess into the chess, thereby rapidly adapting to a new task, namely the migration learning. The transfer learning is to improve the generalization performance of the target model on the target task by transferring the knowledge of the related task under the condition of a small number of label data sets. The transfer learning is widely applied to the field of deep learning.
Considering the mobility of model parameters, there is a lot of work to migrate through parameter migration regularization terms. But these efforts all suffer from severe overfitting problems.
The above background disclosure is only for the purpose of assisting understanding of the concept and technical solution of the present invention and does not necessarily belong to the prior art of the present patent application, and should not be used for evaluating the novelty and inventive step of the present application in the case that there is no clear evidence that the above content is disclosed at the filing date of the present patent application.
Disclosure of Invention
In order to solve the existing problems, the invention provides a migration ensemble learning method, a terminal device and a computer readable storage medium.
In order to solve the above problems, the technical solution adopted by the present invention is as follows:
a migration ensemble learning method, comprising the steps of: s1: uniformly sampling at least 3 different total migratable degrees of the source models from 0 to 1, wherein each total migratable degree of the source models corresponds to one individual learner; s2: for each individual learner, determining an optimal migration strategy based on the total migratable extent of the corresponding source model; s3: migrating a source model to an individual learner of a target task sequentially through a parameter migration regular term and a task loss function based on the optimal migration strategy corresponding to each individual learner; s4: and obtaining an integrated model in an average mode based on the individual learners of each target task.
Preferably, each of said source models has a total migratable extent cmCorresponding to one individual learner fm(·;ωt,m) And sampling M total migratable degrees of the source models, wherein the total migratable degree of the source models is
Figure BDA0002601718100000021
Preferably, the estimating the optimal migration strategy according to the total migratable extent of each source model includes:
Figure BDA0002601718100000022
Figure BDA0002601718100000023
Figure BDA0002601718100000024
wherein ,
Figure BDA0002601718100000025
is the training set of the target task, c is the total migratable extent of the source model, λ is the migration strategy, LTIs a loss function of the target task, ωs,ωtParameters, λ, of the individual learners representing the source model parameters and the target task, respectivelylIs a migration policy per layer, NlIs the number of parameters of the l-th layer, and gamma is a normalizationA constant.
Preferably, the estimating of the optimal migration strategy through the total migratable extent of each source model includes:
Figure BDA0002601718100000026
wherein ,σi,jIs a loss function L of the target taskTSecond derivative of jth model parameter at ith layer; u is an auxiliary variable by which the migration strategy λ is made to satisfy the following condition:
Figure BDA0002601718100000027
in particular, the method of manufacturing a semiconductor device,
Figure BDA0002601718100000031
preferably, the parameter migration regularization term is:
Figure BDA0002601718100000032
the loss function of the target model is:
Figure BDA0002601718100000033
preferably, the optimal migration strategy λ and the pre-training model parameter ω corresponding to each individual learner are based onsTarget model parameters omega of the target task through a complete loss function LtOptimizing to obtain the individual learner f (·; omega)t)。
Preferably, M migration policies based on the selection
Figure BDA0002601718100000034
Obtaining M in turnThe individual learning device
Figure BDA0002601718100000035
wherein ,fmDenotes the m-th learner, ωt,mRepresenting the corresponding model parameters.
Preferably, M of said individual learners are trained
Figure BDA0002601718100000036
And obtaining an integrated model F ═ Sigma by an averaging modemfm
The invention also provides a migration ensemble learning terminal device, which comprises a memory, a processor and a computer program stored in the memory and capable of running on the processor, wherein the processor executes the computer program to realize the steps of any one of the methods.
The invention further provides a computer-readable storage medium having stored thereon a computer program which, when executed by a processor, carries out the steps of the method as set forth in any of the above.
The invention has the beneficial effects that: the method comprises the steps of sampling at least 3 different source model total migratable degrees from 0 to 1 to ensure the difference between individual learners, estimating a corresponding optimal migration strategy according to the different source model total migratable degrees, obtaining a corresponding number of individual learners through a migration strategy and a loss function, and obtaining an integrated model in an averaging mode. By utilizing the effectiveness of ensemble learning in relieving overfitting, the estimation of the migration strategy is more stable and efficient through reasonable assumption, and finally the generalization capability of the ensemble model is greatly improved.
Drawings
Fig. 1 is a schematic diagram of a migration ensemble learning method according to an embodiment of the present invention.
Detailed Description
In order to make the technical problems, technical solutions and advantageous effects to be solved by the embodiments of the present invention more clearly apparent, the present invention is further described in detail below with reference to the accompanying drawings and the embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
It will be understood that when an element is referred to as being "secured to" or "disposed on" another element, it can be directly on the other element or be indirectly on the other element. When an element is referred to as being "connected to" another element, it can be directly connected to the other element or be indirectly connected to the other element. In addition, the connection may be for either a fixing function or a circuit connection function.
It is to be understood that the terms "length," "width," "upper," "lower," "front," "rear," "left," "right," "vertical," "horizontal," "top," "bottom," "inner," "outer," and the like are used in an orientation or positional relationship indicated in the drawings for convenience in describing the embodiments of the present invention and to simplify the description, and are not intended to indicate or imply that the referenced device or element must have a particular orientation, be constructed in a particular orientation, and be in any way limiting of the present invention.
Furthermore, the terms "first", "second" and "first" are used for descriptive purposes only and are not to be construed as indicating or implying relative importance or implicitly indicating the number of technical features indicated. Thus, a feature defined as "first" or "second" may explicitly or implicitly include one or more of that feature. In the description of the embodiments of the present invention, "a plurality" means two or more unless specifically limited otherwise.
In the invention, the source model refers to a pre-training model of a related task; the integrated model is a strong learner formed by training a plurality of individual learners and finally through a certain combination strategy. Related tasks refer to tasks related to the target task. In one particular embodiment, the target task is cat and dog discrimination, and the related task may be tiger and wolf discrimination. In practical applications, there are typically a large number of relevant tasks gathered before the target task is resolved, which may have been previously utilized. Now for the target task, in addition to the data of the target task, the previous related task knowledge can be salvaged.
As shown in fig. 1, the present invention provides a migration ensemble learning method, comprising the following steps:
s1: uniformly sampling at least 3 different total migratable degrees of the source models from 0 to 1, wherein each total migratable degree of the source models corresponds to one individual learner;
s2: for each individual learner, determining an optimal migration strategy based on the total migratable extent of the corresponding source model;
s3: migrating a source model to an individual learner of a target task sequentially through a parameter migration regular term and a task loss function based on the optimal migration strategy corresponding to each individual learner;
s4: and obtaining an integrated model in an average mode based on the individual learners of each target task.
The invention considers the effectiveness of ensemble learning in relieving overfitting through research, and can be introduced into transfer learning. Considering that the superiority of individual learning and the difference of individual learners are important factors for relieving overfitting after integration, the invention ensures the difference between the individual learners by sampling at least 3 different total migratable degrees of source models from 0 to 1, then estimates the corresponding optimal migration strategy according to the total migratable degrees of the different source models, finally can obtain the corresponding number of individual learners through the migration strategy and the loss function, and obtains the integrated model through an average mode.
According to the invention, according to the migratability of the relevant task pre-training model, namely the source model, the generalization performance of the target model of the target task can be improved by migrating the relevant task pre-training model, and then according to the effectiveness of the ensemble learning in relieving the overfitting, the ensemble learning can be introduced into the migration learning, so that the model performance is further improved.
The invention is oriented to a migration ensemble learning method based on model parameters, utilizes the effectiveness of ensemble learning in relieving overfitting, further enables the estimation of a migration strategy to be more stable and efficient through reasonable assumption, and finally greatly improves the generalization capability of an ensemble model.
In an embodiment of the present invention, each of the source models has a total migratable extent cmCorresponding to one individual learner fm(·;ωt,m) And sampling M total migratable degrees of the source models, wherein the total migratable degree of the source models is
Figure BDA0002601718100000051
Since the over-fitting can be relieved by the difference between the individual learners, the difference between the source model and the target model of the target task can be ensured by changing the total migratable degree of the source model, so that the over-fitting is relieved to the maximum extent. When the source model f (·; ω)s) And a target model f (-) of the target task; omegat) When the output of (a) is close, the total migratability of the source model can be considered high. Further, when the source model and the target model of the target task have the same structure, the source model parameter ω can be considered as the source model parameter ωsAnd target model parameters omegatWhen the similarity is high, the model f (·; omega) is pre-traineds) And model f (·; omegat) The output of (c) is close, so the source model is always migratable to a high degree. Controlling source model parameters omega by a migration strategy lambdasAnd target model parameters omegatThe similarity of (c). It is thus possible to use a function of the migration strategy lambda as a proxy value for which the total migratability as source model is high. Considering the influence of different layer parameters on the total migratable degree of the source model, the model migratable degree
Figure BDA0002601718100000061
wherein NlIs the number of parameters for the l-th layer and gamma is a normalization constant. Before estimating M migration strategies, the invention uniformly samples M different total model migratability degrees from 0 to 1
Figure BDA0002601718100000062
Wherein each cmCorresponding to an individual learner fm(·;ωt,m) And selecting the value of M according to the actual storage space and the actual operation resource, wherein M is at least more than 3.
The method for estimating the optimal migration strategy through the total migratable degree of each source model comprises the following steps:
Figure BDA0002601718100000063
Figure BDA0002601718100000064
Figure BDA0002601718100000065
wherein ,
Figure BDA0002601718100000066
is the training set of the target task, c is the total migratable extent of the source model, λ is the migration strategy, LTIs a loss function of the target task, ωs,ωtParameters, λ, of the individual learners representing the source model parameters and the target task, respectivelylIs a migration policy per layer, NlIs the number of parameters of the l-th layer, gamma is a normalization constant,
Figure BDA0002601718100000067
are the individual learner parameters estimated by the integrity loss function.
It will be appreciated that the above description, for simplicity of calculation, will refer to the total migratable extent c of the source modelmThe specific meaning of c is the same, here merely to simplify the formulation in the case of clearance.
Further, estimating an optimal migration strategy through the total migratable extent of each source model, including:
Figure BDA0002601718100000068
wherein ,σi,jIs a loss function L of the target taskTSecond derivative of jth model parameter at ith layer; u is an auxiliary variable by which the migration strategy λ is made to satisfy the following condition:
Figure BDA0002601718100000069
in particular, the method of manufacturing a semiconductor device,
Figure BDA0002601718100000071
by a reasonable approximation of two-stage optimization, u is estimated first and then the specific migration strategy λ can be estimated by the above equation.
Furthermore, M corresponding migration strategies can be obtained by transforming the total model migration degree.
According to the model parameter mobility, a migration strategy lambda of each layer can be definedlThen, the regular term Ω and the migration strategy λ ═ λ are migrated through the parameters1,λ2,...,λL]Migration of knowledge from the source model can be achieved.
The parameter migration regularization term is:
Figure BDA0002601718100000072
considering that the individual learners need to be trained on the target task at the same time to improve the performance of the target model on the target task, a target task loss function L is addedTThe loss function of the target model is:
Figure BDA0002601718100000073
learning based on each of the individual learnersCorresponding optimal migration strategy lambda and pre-training model parameter omegasTarget model parameters omega of the target task through a complete loss function LtOptimizing to obtain the individual learner f (·; omega)t)。
Further, based on the M migration policies selected
Figure BDA0002601718100000074
Sequentially obtaining M individual learners
Figure BDA0002601718100000075
wherein ,fmDenotes the m-th learner, ωt,mRepresenting the corresponding model parameters.
Training M individual learners
Figure BDA0002601718100000076
And obtaining an integrated model F ═ Sigma by an averaging modemfm
To verify the effectiveness of the method of the present invention, the method of the present invention was applied to bird detection systems. The wild bird detection system aims at distinguishing the categories of birds in the wild environment, and the core problem is how to distinguish the categories of the birds in the image from a complex background through a small number of bird training sets. Because there are only a few bird training sets, the prior art cannot get a good model through a classical deep learning network. The method of the invention was applied to this system, which in particular contained 200 birds, a training set size of 5994 and a test set size of 5794, this data being provided by the open data set CUBS. Compared to mainstream methods such as finenet, FitNet, ABdistill, L2SP, cross stick, BAN, SE, Bagging, the method of the present invention exceeds the related mainstream methods by up to about 1% accuracy, which can prove the effectiveness of the patent. The effectiveness of the method of the present invention in mitigating overfitting and improving model performance is demonstrated.
An embodiment of the present application further provides a control apparatus, including a processor and a storage medium for storing a computer program; wherein a processor is adapted to perform at least the method as described above when executing the computer program.
Embodiments of the present application also provide a storage medium for storing a computer program, which when executed performs at least the method described above.
Embodiments of the present application further provide a processor, where the processor executes a computer program to perform at least the method described above.
The storage medium may be implemented by any type of volatile or non-volatile storage device, or combination thereof. Among them, the nonvolatile Memory may be a Read Only Memory (ROM), a Programmable Read Only Memory (PROM), an erasable Programmable Read-Only Memory (EPROM), an electrically erasable Programmable Read-Only Memory (EEPROM), a magnetic random Access Memory (FRAM), a Flash Memory (Flash Memory), a magnetic surface Memory, an optical Disc, or a Compact Disc Read-Only Memory (CD-ROM); the magnetic surface storage may be disk storage or tape storage. Volatile Memory can be Random Access Memory (RAM), which acts as external cache Memory. By way of illustration and not limitation, many forms of RAM are available, such as Static Random Access Memory (SRAM), Synchronous Static Random Access Memory (SSRAM), Dynamic Random Access Memory (DRAM), Synchronous Dynamic Random Access Memory (SDRAM), Double Data rate Synchronous Dynamic Random Access Memory (DDRSDRAM, Double Data rate Synchronous Dynamic Random Access Memory), Enhanced Synchronous Dynamic Random Access Memory (ESDRAM, Enhanced Synchronous Dynamic Random Access Memory), Synchronous link Dynamic Random Access Memory (SLDRAM, Synchronous Dynamic Random Access Memory (DRAM), Direct Memory (DRM, Random Access Memory). The storage media described in connection with the embodiments of the invention are intended to comprise, without being limited to, these and any other suitable types of memory.
In the several embodiments provided in the present application, it should be understood that the disclosed system and method may be implemented in other ways. The above-described device embodiments are merely illustrative, for example, the division of the unit is only a logical functional division, and there may be other division ways in actual implementation, such as: multiple units or components may be combined, or may be integrated into another system, or some features may be omitted, or not implemented. In addition, the coupling, direct coupling or communication connection between the components shown or discussed may be through some interfaces, and the indirect coupling or communication connection between the devices or units may be electrical, mechanical or other forms.
The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed on a plurality of network units; some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, all the functional units in the embodiments of the present invention may be integrated into one processing unit, or each unit may be separately regarded as one unit, or two or more units may be integrated into one unit; the integrated unit can be realized in a form of hardware, or in a form of hardware plus a software functional unit.
Those of ordinary skill in the art will understand that: all or part of the steps for implementing the method embodiments may be implemented by hardware related to program instructions, and the program may be stored in a computer readable storage medium, and when executed, the program performs the steps including the method embodiments; and the aforementioned storage medium includes: a mobile storage device, a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and other various media capable of storing program codes.
Alternatively, the integrated unit of the present invention may be stored in a computer-readable storage medium if it is implemented in the form of a software functional module and sold or used as a separate product. Based on such understanding, the technical solutions of the embodiments of the present invention may be essentially implemented or a part contributing to the prior art may be embodied in the form of a software product, which is stored in a storage medium and includes several instructions for causing a computer device (which may be a personal computer, a server, or a network device) to execute all or part of the methods described in the embodiments of the present invention. And the aforementioned storage medium includes: a removable storage device, a ROM, a RAM, a magnetic or optical disk, or various other media that can store program code.
The methods disclosed in the several method embodiments provided in the present application may be combined arbitrarily without conflict to obtain new method embodiments.
Features disclosed in several of the product embodiments provided in the present application may be combined in any combination to yield new product embodiments without conflict.
The features disclosed in the several method or apparatus embodiments provided in the present application may be combined arbitrarily, without conflict, to arrive at new method embodiments or apparatus embodiments.
The foregoing is a more detailed description of the invention in connection with specific preferred embodiments and it is not intended that the invention be limited to these specific details. For those skilled in the art to which the invention pertains, several equivalent substitutions or obvious modifications can be made without departing from the spirit of the invention, and all the properties or uses are considered to be within the scope of the invention.

Claims (10)

1. A migration ensemble learning method, characterized by comprising the steps of:
s1: uniformly sampling at least 3 different total migratable degrees of the source models from 0 to 1, wherein each total migratable degree of the source models corresponds to one individual learner;
s2: for each individual learner, determining an optimal migration strategy based on the total migratable extent of the corresponding source model;
s3: migrating a source model to an individual learner of a target task sequentially through a parameter migration regular term and a task loss function based on the optimal migration strategy corresponding to each individual learner;
s4: and obtaining an integrated model in an average mode based on the individual learners of each target task.
2. The migration ensemble learning method of claim 1, wherein each of said source models has a total migratable extent cmCorresponding to one individual learner fm(·;ωt,m) And sampling M total migratable degrees of the source models, wherein the total migratable degree of the source models is
Figure FDA0002601718090000011
3. The migration ensemble learning method of claim 2, wherein estimating an optimal migration policy by a total migratable extent of each of said source models comprises:
Figure FDA0002601718090000012
Figure FDA0002601718090000013
Figure FDA0002601718090000014
wherein ,
Figure FDA0002601718090000015
is the training set of the target task, c is the total migratable extent of the source model, λ is the migration strategy, LTIs a loss function of the target task, ωs,ωtParameters, λ, of the individual learners representing the source model parameters and the target task, respectivelylIs a migration policy per layer, NlIs the number of parameters for the l-th layer and gamma is a normalization constant.
4. The migration ensemble learning method of claim 3, wherein estimating an optimal migration strategy by a total migratable extent of each of said source models comprises:
Figure FDA0002601718090000016
wherein ,σi,jIs a loss function L of the target taskTSecond derivative of jth model parameter at ith layer; u is an auxiliary variable by which the migration strategy λ is made to satisfy the following condition:
Figure FDA0002601718090000021
in particular, the method of manufacturing a semiconductor device,
Figure FDA0002601718090000022
5. the migration ensemble learning method of claim 4, wherein said parameter migration regularization term is:
Figure FDA0002601718090000023
the loss function of the target model is:
Figure FDA0002601718090000024
6. the migration ensemble learning method according to claim 5, wherein the optimal migration strategy λ and pre-training model parameter ω are based on the optimal migration strategy λ and pre-training model parameter ω corresponding to each of the individual learnerssTarget model parameters omega of the target task through a complete loss function LtOptimizing to obtain the individual learner f (·; omega)t)。
7. The migration ensemble learning method of claim 6, wherein the M migration strategies based on the selection
Figure FDA0002601718090000025
Sequentially obtaining M individual learners
Figure FDA0002601718090000026
wherein ,fmDenotes the m-th learner, ωt,mRepresenting the corresponding model parameters.
8. The method of migratory ensemble learning of claim 7, wherein M of said individual learners are trained
Figure FDA0002601718090000027
And obtaining an integrated model F ═ Sigma by an averaging modemfm
9. A migration ensemble learning terminal device comprising a memory, a processor and a computer program stored in said memory and executable on said processor, characterized in that said processor, when executing said computer program, implements the steps of the method according to any of claims 1-8.
10. A computer-readable storage medium, in which a computer program is stored which, when being executed by a processor, carries out the steps of the method according to any one of claims 1 to 8.
CN202010726052.2A 2020-07-24 2020-07-24 Migration integrated learning method, terminal device and computer readable storage medium Active CN111950736B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010726052.2A CN111950736B (en) 2020-07-24 2020-07-24 Migration integrated learning method, terminal device and computer readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010726052.2A CN111950736B (en) 2020-07-24 2020-07-24 Migration integrated learning method, terminal device and computer readable storage medium

Publications (2)

Publication Number Publication Date
CN111950736A true CN111950736A (en) 2020-11-17
CN111950736B CN111950736B (en) 2023-09-19

Family

ID=73338081

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010726052.2A Active CN111950736B (en) 2020-07-24 2020-07-24 Migration integrated learning method, terminal device and computer readable storage medium

Country Status (1)

Country Link
CN (1) CN111950736B (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105300693A (en) * 2015-09-25 2016-02-03 东南大学 Bearing fault diagnosis method based on transfer learning
CN106803124A (en) * 2017-01-21 2017-06-06 中国海洋大学 Field migration extreme learning machine method based on manifold canonical and norm canonical
CN110032646A (en) * 2019-05-08 2019-07-19 山西财经大学 The cross-domain texts sensibility classification method of combination learning is adapted to based on multi-source field
CN110647920A (en) * 2019-08-29 2020-01-03 北京百度网讯科技有限公司 Transfer learning method and device in machine learning, equipment and readable medium
US20200012923A1 (en) * 2016-10-06 2020-01-09 Siemens Aktiengesellschaft Computer device for training a deep neural network

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105300693A (en) * 2015-09-25 2016-02-03 东南大学 Bearing fault diagnosis method based on transfer learning
US20200012923A1 (en) * 2016-10-06 2020-01-09 Siemens Aktiengesellschaft Computer device for training a deep neural network
CN106803124A (en) * 2017-01-21 2017-06-06 中国海洋大学 Field migration extreme learning machine method based on manifold canonical and norm canonical
CN110032646A (en) * 2019-05-08 2019-07-19 山西财经大学 The cross-domain texts sensibility classification method of combination learning is adapted to based on multi-source field
CN110647920A (en) * 2019-08-29 2020-01-03 北京百度网讯科技有限公司 Transfer learning method and device in machine learning, equipment and readable medium

Also Published As

Publication number Publication date
CN111950736B (en) 2023-09-19

Similar Documents

Publication Publication Date Title
CN110892417B (en) Asynchronous agent with learning coaches and structurally modifying deep neural networks without degrading performance
WO2020259582A1 (en) Neural network model training method and apparatus, and electronic device
EP3779774B1 (en) Training method for image semantic segmentation model and server
Avgar et al. Space‐use behaviour of woodland caribou based on a cognitive movement model
US20220414464A1 (en) Method and server for federated machine learning
CN110347932B (en) Cross-network user alignment method based on deep learning
CN110674323B (en) Unsupervised cross-modal Hash retrieval method and system based on virtual label regression
CN110377587B (en) Migration data determination method, device, equipment and medium based on machine learning
CN111428874A (en) Wind control method, electronic device and computer readable storage medium
TW202004559A (en) Feature interpretation method and device for GBDT model
CN110866124A (en) Medical knowledge graph fusion method and device based on multiple data sources
KR20210032140A (en) Method and apparatus for performing pruning of neural network
US11481810B2 (en) Generating and utilizing machine-learning models to create target audiences with customized auto-tunable reach and accuracy
CN111261286A (en) Auxiliary diagnosis model construction method, diagnosis method, device, equipment and medium
CN113257361B (en) Method, device and equipment for realizing self-adaptive protein prediction framework
CN115019106A (en) Robust unsupervised domain self-adaptive image classification method and device based on anti-distillation
US20220121924A1 (en) Configuring a neural network using smoothing splines
US20220083907A1 (en) Data generation and annotation for machine learning
CN111950736A (en) Migration ensemble learning method, terminal device, and computer-readable storage medium
CN113222014A (en) Image classification model training method and device, computer equipment and storage medium
US20230325373A1 (en) Machine-learning based automated document integration into genealogical trees
CN117010480A (en) Model training method, device, equipment, storage medium and program product
CN115689981A (en) Lung image detection method and device based on information fusion and storage medium
CN115346084A (en) Sample processing method, sample processing apparatus, electronic device, storage medium, and program product
Rotter Relevance feedback based on n-tuplewise comparison and the ELECTRE methodology and an application in content-based image retrieval

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant