CN112116104A - Method, apparatus, medium, and electronic device for automatically integrating machine learning - Google Patents

Method, apparatus, medium, and electronic device for automatically integrating machine learning Download PDF

Info

Publication number
CN112116104A
CN112116104A CN202010982700.0A CN202010982700A CN112116104A CN 112116104 A CN112116104 A CN 112116104A CN 202010982700 A CN202010982700 A CN 202010982700A CN 112116104 A CN112116104 A CN 112116104A
Authority
CN
China
Prior art keywords
integration
machine learning
model
parameter
hyper
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010982700.0A
Other languages
Chinese (zh)
Other versions
CN112116104B (en
Inventor
夏晓华
王美青
吕军
程建波
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
JD Digital Technology Holdings Co Ltd
Original Assignee
JD Digital Technology Holdings Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by JD Digital Technology Holdings Co Ltd filed Critical JD Digital Technology Holdings Co Ltd
Priority to CN202010982700.0A priority Critical patent/CN112116104B/en
Publication of CN112116104A publication Critical patent/CN112116104A/en
Application granted granted Critical
Publication of CN112116104B publication Critical patent/CN112116104B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • G06N20/20Ensemble learning

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Medical Informatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Artificial Intelligence (AREA)
  • Feedback Control In General (AREA)
  • Machine Translation (AREA)

Abstract

The embodiment of the invention provides a method, a device, a medium and electronic equipment for automatically integrating machine learning, wherein the method comprises the following steps: step S1, acquiring at least two pre-configured machine learning models, a hyper-parameter search space, an integration algorithm and an integration algorithm parameter set; step S2, inputting a first hyper-parameter selected in the range of the hyper-parameter search space into each machine learning model; step S3, integrating the at least two machine learning models according to the integration algorithm and the first integration algorithm parameters to generate a first integration model; step S4, training the first integrated model and scoring the model to obtain a scoring result; and step S5, taking the step S2, the step S3 and the step S4 which are executed in sequence as a loop, and after the loop is repeatedly executed, determining a target integration model according to the grading result of each loop. The technical scheme of the embodiment of the invention can obtain an integrated model with better overall quality.

Description

Method, apparatus, medium, and electronic device for automatically integrating machine learning
Technical Field
The invention relates to the technical field of computers, in particular to a method and a device for automatically integrating machine learning, a computer readable storage medium and electronic equipment.
Background
Machine Learning (Machine Learning) is a multi-field cross subject, and relates to a plurality of subjects such as probability theory, statistics, approximation theory, convex analysis, algorithm complexity theory and the like. Machine learning specializes in studying how computers simulate or implement human learning behaviors to acquire new knowledge or skills and reorganize existing knowledge structures to improve their performance. Machine learning is the core of artificial intelligence, is the fundamental approach for computers to have intelligence, and thus is also increasingly widely used in various business scenarios.
By processing and selecting the service data, processing and selecting the characteristics, selecting the model and adjusting the parameters, the optimal machine learning model can be obtained. This process involves a great deal of repetitive and empirical labor.
The automatic machine learning technology automatically generates a machine learning model through means of automatic characteristic engineering, automatic model selection, automatic parameter adjustment optimization and the like, and reduces the labor amount of algorithm personnel.
In order to improve the effect of the machine learning model, the multiple models can be integrated by methods such as ensemble learning from multiple aspects such as accuracy, precision and recall rate to obtain an integrated model with higher effect score.
Ensemble learning is a technique that uses multiple compatible learning algorithms/models to perform a single task in order to achieve better predictive performance.
The prior art typically optimizes individual models over-parameters by automatic machine learning. The ensemble learning is generally performed after the basic model is optimized, and parameters of the ensemble learning also need to be adjusted and optimized. When the processes of carrying out hyper-parameter optimization on a single model and carrying out parameter optimization on ensemble learning are carried out in series according to the sequence, the optimization of adjusting and optimizing by stages is actually carried out, and although the obtained final model is adjusted and optimized in each stage, the final integral model is not necessarily the optimal model.
It is to be noted that the information disclosed in the above background section is only for enhancement of understanding of the background of the present invention and therefore may include information that does not constitute prior art known to a person of ordinary skill in the art.
Disclosure of Invention
Embodiments of the present invention provide a method, an apparatus, a computer-readable storage medium, and an electronic device for automatically integrating machine learning, so as to obtain an integrated model with overall superiority to at least a certain extent.
Additional features and advantages of the invention will be set forth in the detailed description which follows, or may be learned by practice of the invention.
According to a first aspect of embodiments of the present invention, there is provided a method of automatically integrating machine learning, the method comprising: step S1, acquiring at least two pre-configured machine learning models, a hyper-parameter search space of each machine learning model, an integration algorithm and an integration algorithm parameter group of the integration algorithm, wherein the integration algorithm parameter group comprises at least one integration algorithm parameter; step S2, inputting a first hyper-parameter selected in the range of the hyper-parameter search space into each machine learning model; step S3, integrating the at least two machine learning models according to the integration algorithm and a first integration algorithm parameter in the integration algorithm parameter group to generate a first integration model; step S4, training the first integrated model and scoring the model to obtain a scoring result; and step S5, taking the step S2, the step S3 and the step S4 which are executed in sequence as a loop, and after N loops are repeatedly executed, determining a target integration model according to the grading result of each loop, wherein N is a positive integer.
In some embodiments, the step S2 includes: integrating the at least two machine learning models according to any one of the following integration algorithms: stacking integration algorithms, bagging integration algorithms, and lifting integration algorithms.
In some embodiments, before the step S1, the method further comprises: generating a configuration file in a json format, wherein the configuration file comprises the at least two machine learning models, a hyper-parameter search space of each machine learning model, at least two integration algorithms and integration algorithm parameters of the integration algorithms.
In some embodiments, the determining the target integration model according to the scoring result of each cycle includes: and determining a target integration model according to the grading result of each cycle and the cycle execution times.
In some embodiments, after the step S4, the method further includes: and selecting the hyper-parameters and the integration algorithm parameters of the next cycle according to the grading result and the optimization algorithm.
In some embodiments, the machine learning model is derived from any one of the following machine learning frameworks: tensorflow symbolic math system, a pytorch deep learning framework, a sklern machine learning language library.
In some embodiments, the machine learning model has weight parameters, the method further comprising: and determining the weight of each machine learning model in the integration process according to the weight parameter of each machine learning model.
According to a second aspect of embodiments of the present invention, there is provided an apparatus for automatically integrating machine learning, the apparatus comprising: the system comprises an acquisition unit, a processing unit and a processing unit, wherein the acquisition unit is used for acquiring at least two pre-configured machine learning models, a hyper-parameter search space of each machine learning model, an integration algorithm and an integration algorithm parameter group of the integration algorithm, and the integration algorithm parameter group comprises at least one integration algorithm parameter; a parameter input unit configured to input a first hyper-parameter selected within a range of the hyper-parameter search space into each of the machine learning models; the integration unit is used for integrating the at least two machine learning models according to an integration algorithm and a first integration algorithm parameter in the integration algorithm parameter group to generate a first integration model; the training scoring unit is used for training and scoring the first integrated model to obtain a scoring result; and the preference unit is used for determining a target integrated model according to the grading result of each cycle after the input unit, the integration unit and the training grading unit repeat N times of cycles of input, integration, training and grading in sequence, wherein N is a positive integer.
In some embodiments, the integrated unit is further configured to: integrating the at least two machine learning models according to any one of the following integration algorithms: stacking integration algorithms, bagging integration algorithms, and lifting integration algorithms.
In some embodiments, the apparatus further comprises: a configuration unit, configured to generate a configuration file in a json format, where the configuration file includes the at least two machine learning models, a hyper-parameter search space of each machine learning model, at least two integration algorithms, and an integration algorithm parameter set of each integration algorithm.
In some embodiments, the target integration model includes an optimal hyper-parameter and an optimal integration algorithm parameter, and the preference unit is further configured to determine the target integration model according to the scoring result of each loop and the number of times of loop execution.
In some embodiments, the apparatus further comprises: and the selecting unit is used for selecting the hyper-parameters and the integrated algorithm parameters of the next cycle according to the grading result and the optimization algorithm.
In some embodiments, the machine learning models have weight parameters, and the apparatus further comprises a determining unit for determining the weight of each of the machine learning models in the integration process according to the weight parameters of each of the machine learning models.
According to a third aspect of embodiments of the present invention, there is provided a computer readable storage medium having stored thereon a computer program which, when executed by a processor, implements the method of automatic integrated machine learning as described in the first aspect of the embodiments above.
According to a fourth aspect of embodiments of the present invention, there is provided an electronic apparatus, including: one or several processors; storage means for storing one or several programs which, when executed by the one or several processors, cause the one or several processors to carry out the method of automatic integrated machine learning as described in the first aspect of the embodiments above.
The technical scheme provided by the embodiment of the invention has the following beneficial effects:
in the technical solutions provided in some embodiments of the present invention, an overall optimal integrated model can be obtained by integrating the machine learning model, training and scoring the model, repeating the cycle for a plurality of times, and determining a target integrated model according to the scoring result.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the invention, as claimed.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the invention and together with the description, serve to explain the principles of the invention. It is obvious that the drawings in the following description are only some embodiments of the invention, and that for a person skilled in the art, other drawings can be derived from them without inventive effort. In the drawings:
FIG. 1 schematically illustrates a flow diagram of a method of automatically integrating machine learning, in accordance with one embodiment of the present invention;
FIG. 2 schematically illustrates a flow diagram of a method of automatically integrating machine learning, in accordance with another embodiment of the present invention;
FIG. 3 schematically illustrates a block diagram of an apparatus for automated integrated machine learning, in accordance with one embodiment of the present invention;
FIG. 4 illustrates a schematic structural diagram of a computer system suitable for use with the electronic device to implement an embodiment of the invention.
Detailed Description
Example embodiments will now be described more fully with reference to the accompanying drawings. Example embodiments may, however, be embodied in many different forms and should not be construed as limited to the examples set forth herein; rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the concept of example embodiments to those skilled in the art.
Furthermore, the described features, structures, or characteristics may be combined in any suitable manner in one or more embodiments. In the following description, numerous specific details are provided to provide a thorough understanding of embodiments of the invention. One skilled in the relevant art will recognize, however, that the invention may be practiced without one or more of the specific details, or with other methods, components, devices, steps, and so forth. In other instances, well-known methods, devices, implementations or operations have not been shown or described in detail to avoid obscuring aspects of the invention.
The block diagrams shown in the figures are functional entities only and do not necessarily correspond to physically separate entities. I.e. these functional entities may be implemented in the form of software, or in one or several hardware modules or integrated circuits, or in different networks and/or processor means and/or microcontroller means.
The flow charts shown in the drawings are merely illustrative and do not necessarily include all of the contents and operations/steps, nor do they necessarily have to be performed in the order described. For example, some operations/steps may be decomposed, and some operations/steps may be combined or partially combined, so that the actual execution sequence may be changed according to the actual situation.
Fig. 1 is a flow chart illustrating a method for automatically integrating machine learning according to an embodiment of the present invention. The method provided by the embodiment of the invention can be executed by any electronic equipment with computer processing capability, such as a terminal device and/or a server. As shown in fig. 1, the method of automatically integrating machine learning includes:
and step S1, acquiring at least two pre-configured machine learning models, a hyper-parameter search space of each machine learning model, an integration algorithm and an integration algorithm parameter group of the integration algorithm, wherein the integration algorithm group comprises at least one integration algorithm parameter.
Step S2, inputting the first hyper-parameter selected in the range of the hyper-parameter search space into each machine learning model.
And step S3, integrating at least two machine learning models according to the integration algorithm and the first integration algorithm parameter in the integration algorithm parameter group to generate a first integration model.
And step S4, training the first integrated model and scoring the model to obtain a scoring result.
And step S5, taking the step S2, the step S3 and the step S4 which are sequentially executed as a loop, repeatedly executing N loops, and determining the target integrated model according to the scoring result of each loop, wherein N is a positive integer.
In the technical scheme of the embodiment of the invention, a first hyper-parameter is selected in the range of a hyper-parameter search space, a first integration algorithm parameter is selected from an integration algorithm parameter group, a plurality of machine learning models are subjected to parameter input and integration according to the first hyper-parameter and the first integration algorithm parameter, the obtained integration models are subjected to model scoring to form a cycle, and after the cycle is executed for a plurality of times, the overall optimal integration model can be determined as a target integration model according to the scoring result of each cycle.
In the embodiment of the invention, different integration algorithms can be selected in different cycles.
The technical scheme of the embodiment of the invention combines the super-parameter tuning of the machine learning model with the integrated algorithm selection, the integrated learning parameter selection and other steps, and the like to operate, and performs integrated learning at the same time in the automatic machine learning stage to obtain the final overall optimization model, thereby reducing the labor amount of algorithm personnel, and simultaneously obtaining the final integrated model which is the result of overall scoring optimization, namely the overall optimized integrated model.
With the help of the technical scheme of the embodiment of the invention, an algorithm worker can easily perform automatic ensemble learning training after compiling the hyper-parameter search space file according to a certain format rule, and finally generate an optimized ensemble model. In the whole process, a large amount of repetitive labor of algorithm personnel is avoided, a large amount of time is saved, and the production efficiency is improved.
The implementation of the technical scheme of the embodiment of the invention can depend on the python dependent library. Before step S1, the python library on which the solution of the embodiment of the present invention depends is also automatically installed. And then generating a configuration file in a json format, wherein the configuration file comprises at least two machine learning models, a hyper-parameter search space of each machine learning model, an integration algorithm and an integration algorithm parameter of the integration algorithm.
json (JS object notation) is a lightweight data exchange format. It stores and represents data in a text format that is completely independent of the programming language, based on a subset of ECMAScript (js specification set by the european computer association).
The configuration file is mainly used for defining the configuration of the hyper-parameter search space and the integration model. The configuration file has a certain format rule, and an algorithm worker writes the configuration file according to the format rule so as to load the configuration file when performing automatic machine integration learning.
In the technical scheme of the embodiment of the invention, the configuration file and the optimization algorithm are used for determining the hyper-parameters. The hyper-parameter selected in each cycle is a selection result of an adjustable hyper-parameter search space in a configuration file, the selected hyper-parameter is input into a written model, several models can be integrated according to an integration algorithm and the configuration of parameters of the integration algorithm, and then training and model scoring are performed to complete a cycle.
The hyper-parameters in the first cycle may be chosen randomly or empirically. After the first cycle is completed, the framework selects the next group of hyper-parameters according to the scoring result and the optimization algorithm, starts the next round of training and scoring, and repeats the steps in the cycle until the optimal hyper-parameters and the optimal integration model parameters are searched, so that the target integration model, namely the optimal integration model, is obtained.
The configuration of the parametric search space and the integration model is defined by a json-formatted file, the format of which is represented by the following example:
Figure BDA0002688129140000071
Figure BDA0002688129140000081
Figure BDA0002688129140000091
as seen from the example json configuration file, four machine learning models 0-3 and adjustable hyper-parameter search spaces of the models are arranged in a nested configuration mode, the four machine learning models perform automatic hyper-parameter tuning in the adjustable hyper-parameter spaces, and the last line of configuration is to integrate the four models configured in the front by adopting VotingClassifier (voting classifier).
In practical applications, the number of machine learning models in the configuration file is not limited to four.
In step S2, at least two machine learning models may be integrated according to any one of the following integration algorithms: stacking integration algorithms, bagging integration algorithms, and lifting integration algorithms.
Stacking (Stacking), bagging (bagging), and Boosting are the main methods of three ensemble studies.
In the embodiment of the present invention, the machine learning models have weight parameters, and in step S3, the weight of each machine learning model in the integration process may be determined according to the weight parameters of each machine learning model.
For the json configuration file of the above example, the weight parameter of each machine learning model is selected from the parameter space defined in the configured value, so that tuning of ensemble learning is achieved, and an algorithm person can adopt different ensemble learning according to the algorithm effect.
In step S5, the target integration model may be determined based on the score result of each loop, or may be determined based on the score result of each loop and the number of loop executions.
Here, the target integration model is an optimal integration model, and the target integration model includes an optimal hyper-parameter and an optimal integration algorithm parameter. When the optimal integration model is selected, one way is to consider that the optimal integration model is obtained when the scoring result reaches a set threshold value. In another mode, a curve graph of the grading result and the number of times of cycle execution is constructed, and the highest point of the curve is taken as an optimal integrated model.
After step S4, the hyper-parameters and the parameters of the integrated algorithm of the next cycle are selected according to the scoring result and the optimization algorithm. That is, after step S4, an appropriate hyper-parametric input machine learning model is selected for step S2 based on the scoring results obtained in step S4.
Here, the optimization algorithm may include a bayesian optimization algorithm, a TPE (Tree-structured park Estimator Approach) optimization algorithm, and is not limited thereto.
In one embodiment of the present invention, after 40 times of loop execution, an optimal integration model can be obtained.
In the embodiment of the invention, various parameters of the machine learning model and the deep learning model can be adjusted, so that the performance of the machine learning model is optimal.
In particular, the machine learning model may be derived from any one of the following machine learning frameworks: tensorflow symbolic math system, a pytorch deep learning framework, a sklern machine learning language library.
Among them, the tensflow is a symbolic mathematical system based on dataflow programming, and is widely applied to programming realization of various machine learning (machine learning) algorithms, and the predecessor of the tensflow is the neural network algorithm library distebrief of ***. Pythrch is a popular deep learning framework. skearn is a machine learning Python language library of open source (BSD).
Deep Learning (Deep Learning) is a branch of machine Learning that attempts algorithms that perform high-level abstractions of data using multiple processing layers that contain complex structures or are composed of multiple nonlinear transformations. Deep learning is a characterization learning method in machine learning. The observations of an image may be represented in a number of ways, such as a vector of intensity values for each pixel, or more abstractly as a series of edges, specially shaped regions, and so forth. And tasks are easier to learn from the examples using some specific representation methods. Such as face recognition or facial expression recognition. The benefit of deep learning is that the manual acquisition of features will be replaced by efficient algorithms of unsupervised or semi-supervised feature learning and hierarchical feature extraction.
Since the steps of model hyper-parameter optimization and ensemble learning are all realized by python, machine learning framework libraries supporting all python interfaces, such as popular tensorflow, pyrrch, sklern, are supported. And the integrated algorithm of all python interfaces is supported, and the integrated learning such as stacking, bagging and lifting is realized.
As shown in fig. 2, an automatic integrated machine learning method according to an embodiment of the present invention includes the following steps:
step S601, a plurality of machine learning models which are pre-configured, a hyper-parameter search space of each machine learning model, an integration algorithm and an integration algorithm parameter of the integration algorithm are obtained.
In step S602, one hyper-parameter selected within the range of the hyper-parameter search space is input to each machine learning model.
Step S603, integrating a plurality of machine learning models according to an integration algorithm and an integration algorithm parameter to generate an integrated model.
And step S604, training the integrated model and scoring the model to obtain a scoring result.
And step S605, judging whether the integration model is the optimal integration model according to the grading result. If yes, executing step S607; if the determination result is no, step S606 is executed.
Step S606, selecting a new hyper-parameter according to the scoring result.
And step S607, determining the optimal integration model as the target integration model.
In the method for automatically integrating machine learning provided by the embodiment of the invention, the machine learning model is integrated, then training and model scoring are carried out, the cycle is repeated for multiple times, and the target integration model is determined according to the scoring result, so that the integrally optimal integration model can be obtained.
Embodiments of the apparatus of the present invention are described below that can be used to perform the above-described methods of automated integrated machine learning of the present invention. Referring to fig. 3, an apparatus based on automatic integrated machine learning according to an embodiment of the present invention includes:
an obtaining unit 302, configured to obtain at least two preconfigured machine learning models, a hyper-parameter search space of each machine learning model, an integration algorithm, and an integration algorithm parameter set of the integration algorithm, where the integration algorithm parameter set includes at least one integration algorithm parameter.
A parameter input unit 304 for inputting a first hyper-parameter selected within the range of the hyper-parameter search space into each machine learning model.
An integrating unit 306, configured to integrate the at least two machine learning models according to the integration algorithm and the first integration algorithm parameter of the integration algorithm, so as to generate a first integration model.
And the training scoring unit 308 is configured to train and score the first integrated model to obtain a scoring result.
And the preference unit 310 is configured to determine a target integration model according to a scoring result of each cycle after the input unit, the integration unit, and the training scoring unit repeat N cycles of input, integration, training, and scoring operations in sequence, where N is a positive integer.
The integration unit 306 is further configured to: integrating at least two machine learning models according to any one of the following integration algorithms: stacking integration algorithms, bagging integration algorithms, and lifting integration algorithms.
The device further comprises a configuration unit, which is used for generating a configuration file in a json format, wherein the configuration file comprises at least two machine learning models, a hyper-parameter search space of each machine learning model, at least two integration algorithms and an integration algorithm parameter group of each integration algorithm.
The target integration model includes an optimal hyper-parameter and an optimal integration algorithm parameter, and the optimization unit 310 is further configured to determine the target integration model according to the scoring results of each cycle and the number of times of cycle execution.
The device also comprises a selection unit used for selecting the hyper-parameters and the integrated algorithm parameters of the next cycle according to the grading result and the optimization algorithm.
The machine learning models have weight parameters, and the device further comprises a determining unit for determining the weight of each machine learning model in the integration process according to the weight parameters of each machine learning model.
For details which are not disclosed in the embodiments of the apparatus of the present invention, reference is made to the above-described embodiments of the method for automatic integrated machine learning of the present invention for details which are not disclosed in the embodiments of the apparatus of the present invention, since the respective functional modules of the apparatus for automatic integrated machine learning based on the exemplary embodiments of the present invention correspond to the steps of the above-described exemplary embodiments of the method for automatic integrated machine learning.
In the device based on the automatic integrated machine learning provided by the embodiment of the invention, the machine learning model is integrated firstly, then training and model grading are carried out, the cycle is repeated for multiple times, and the target integrated model is determined according to the grading result, so that the integrally optimal integrated model can be obtained.
Referring now to FIG. 4, a block diagram of a computer system 400 suitable for use with the electronic device implementing an embodiment of the invention is shown. The computer system 400 of the electronic device shown in fig. 4 is only an example, and should not bring any limitation to the function and the scope of use of the embodiments of the present invention.
As shown in fig. 4, the computer system 400 includes a Central Processing Unit (CPU)401 that can perform various appropriate actions and processes in accordance with a program stored in a Read Only Memory (ROM)402 or a program loaded from a storage section 404 into a Random Access Memory (RAM) 403. In the RAM 403, various programs and data necessary for system operation are also stored. The CPU 401, ROM 402, and RAM 403 are connected to each other via a bus 404. An input/output (I/O) interface 405 is also connected to bus 404.
The following components are connected to the I/O interface 405: an input section 406 including a keyboard, a mouse, and the like; an output section 407 including a display device such as a Cathode Ray Tube (CRT), a Liquid Crystal Display (LCD), and the like, and a speaker; a storage section 408 including a hard disk and the like; and a communication section 409 including a network interface card such as a LAN card, a modem, or the like. The communication section 409 performs communication processing via a network such as the internet. A driver 410 is also connected to the I/O interface 405 as needed. A removable medium 411 such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory, or the like is mounted on the drive 410 as necessary, so that a computer program read out therefrom is mounted into the storage section 408 as necessary.
In particular, according to an embodiment of the present invention, the processes described above with reference to the flowcharts may be implemented as computer software programs. For example, embodiments of the invention include a computer program product comprising a computer program embodied on a computer-readable storage medium, the computer program comprising program code for performing the method illustrated by the flow chart. In such an embodiment, the computer program may be downloaded and installed from a network through the communication section 409, and/or installed from the removable medium 411. The above-described functions defined in the system of the present application are executed when the computer program is executed by a Central Processing Unit (CPU) 401.
It should be noted that the computer readable storage medium shown in the present invention can be a computer readable signal medium or a computer readable storage medium or any combination of the two. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples of the computer readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the present invention, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. In the present invention, however, a computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may also be any computer readable storage medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable storage medium may be transmitted using any appropriate medium, including but not limited to: wireless, wire, fiber optic cable, RF, etc., or any suitable combination of the foregoing.
The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams or flowchart illustration, and combinations of blocks in the block diagrams or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The units described in the embodiments of the present invention may be implemented by software, or may be implemented by hardware, and the described units may also be disposed in a processor. Wherein the names of the elements do not in some way constitute a limitation on the elements themselves.
As another aspect, the present application also provides a computer-readable storage medium, which may be included in the electronic device described in the above embodiments; or may exist separately without being assembled into the electronic device. The computer readable storage medium carries one or several programs, which when executed by an electronic device, cause the electronic device to implement the method for automatically integrating machine learning as described in the above embodiments.
For example, the electronic device may implement the following as shown in fig. 1: step S1, acquiring at least two pre-configured machine learning models, a hyper-parameter search space of each machine learning model, an integration algorithm and an integration algorithm parameter group of the integration algorithm, wherein the integration algorithm group comprises at least one integration algorithm parameter; step S2, inputting a first hyper-parameter selected in the range of the hyper-parameter search space into each machine learning model; step S3, integrating the at least two machine learning models according to an integration algorithm and a first integration algorithm parameter in the integration algorithm parameter group to generate a first integration model; step S4, training the first integrated model and scoring the model to obtain a scoring result; and step S5, taking the step S2, the step S3 and the step S4 which are sequentially executed as a loop, repeatedly executing N loops, and determining a target integration model according to the grading result of each loop, wherein N is a positive integer.
As another example, the electronic device may also implement the steps as described in fig. 2.
It should be noted that although in the above detailed description several modules or units of the device for action execution are mentioned, such a division is not mandatory. Indeed, the features and functionality of two or more modules or units described above may be embodied in one module or unit, according to embodiments of the invention. Conversely, the features and functions of one module or unit described above may be further divided into embodiments by several modules or units.
Through the above description of the embodiments, those skilled in the art will readily understand that the exemplary embodiments described herein may be implemented by software, or by software in combination with necessary hardware. Therefore, the technical solution according to the embodiment of the present invention can be embodied in the form of a software product, which can be stored in a non-volatile storage medium (which can be a CD-ROM, a usb disk, a removable hard disk, etc.) or on a network, and includes several instructions to enable a computing device (which can be a personal computer, a server, a touch terminal, or a network device, etc.) to execute the method according to the embodiment of the present invention.
Other embodiments of the invention will be apparent to those skilled in the art from consideration of the specification and practice of the invention disclosed herein. This application is intended to cover any variations, uses, or adaptations of the invention following, in general, the principles of the invention and including such departures from the present disclosure as come within known or customary practice within the art to which the invention pertains. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the invention being indicated by the following claims.
It will be understood that the invention is not limited to the precise arrangements described above and shown in the drawings and that various modifications and changes may be made without departing from the scope thereof. The scope of the invention is limited only by the appended claims.

Claims (10)

1. A method of automatically integrating machine learning, the method comprising the steps of:
step S1, acquiring at least two pre-configured machine learning models, a hyper-parameter search space of each machine learning model, an integration algorithm and an integration algorithm parameter group of the integration algorithm, wherein the integration algorithm parameter group comprises at least one integration algorithm parameter;
step S2, inputting a first hyper-parameter selected in the range of the hyper-parameter search space into each machine learning model;
step S3, integrating the at least two machine learning models according to the integration algorithm and a first integration algorithm parameter in the integration algorithm parameter group to generate a first integration model;
step S4, training the first integrated model and scoring the model to obtain a scoring result;
and step S5, taking the step S2, the step S3 and the step S4 which are executed in sequence as a loop, and after the loop is repeatedly executed, determining a target integration model according to the grading result of each loop.
2. The method according to claim 1, wherein the step S3 includes:
integrating the at least two machine learning models according to any one of the following integration algorithms:
stacking integration algorithms, bagging integration algorithms, and lifting integration algorithms.
3. The method according to claim 1, wherein before the step S1, the method further comprises:
generating a configuration file in a json format, wherein the configuration file comprises the at least two machine learning models, a hyper-parameter search space of each machine learning model, at least two integration algorithms and an integration algorithm parameter set of each integration algorithm.
4. The method of claim 1, wherein the target integration model comprises an optimal hyper-parameter and an optimal integration algorithm parameter, and wherein determining the target integration model according to the scoring results of each of the cycles comprises:
and determining a target integration model according to the grading result of each cycle and the cycle execution times.
5. The method according to claim 1, wherein after the step S4, the method further comprises:
and selecting the hyper-parameters and the integrated algorithm parameters of the next cycle according to the scoring result and the optimization algorithm.
6. The method of claim 1, wherein the machine learning model is derived from any one of the following machine learning frameworks:
tensorflow symbolic math system, a pytorch deep learning framework, a sklern machine learning language library.
7. The method of claim 1, wherein the machine learning model has weight parameters, the method further comprising:
and determining the weight of each machine learning model in the integration process according to the weight parameter of each machine learning model.
8. An apparatus for automatically integrating machine learning, the apparatus comprising:
the system comprises an acquisition unit, a processing unit and a processing unit, wherein the acquisition unit is used for acquiring at least two pre-configured machine learning models, a hyper-parameter search space of each machine learning model, an integration algorithm and an integration algorithm parameter group of the integration algorithm, and the integration algorithm parameter group comprises at least one integration algorithm parameter;
a parameter input unit configured to input a first hyper-parameter selected within a range of the hyper-parameter search space into each of the machine learning models;
the integration unit is used for integrating the at least two machine learning models according to an integration algorithm and a first integration algorithm parameter in the integration algorithm parameter group to generate a first integration model;
the training scoring unit is used for training and scoring the first integrated model to obtain a scoring result;
and the preference unit is used for determining a target integrated model according to the grading result of each cycle after the input unit, the integration unit and the training grading unit repeatedly and sequentially carry out the cycle of input, integration, training and grading operation.
9. A computer-readable storage medium, on which a computer program is stored, which, when being executed by a processor, carries out a method of automatically integrating machine learning according to any one of claims 1 to 7.
10. An electronic device, comprising:
one or several processors;
storage means for storing one or several programs which, when executed by the one or several processors, cause the one or several processors to carry out the method of automated integrated machine learning according to any one of claims 1 to 7.
CN202010982700.0A 2020-09-17 2020-09-17 Method, device, medium and electronic equipment for automatically integrating machine learning Active CN112116104B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010982700.0A CN112116104B (en) 2020-09-17 2020-09-17 Method, device, medium and electronic equipment for automatically integrating machine learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010982700.0A CN112116104B (en) 2020-09-17 2020-09-17 Method, device, medium and electronic equipment for automatically integrating machine learning

Publications (2)

Publication Number Publication Date
CN112116104A true CN112116104A (en) 2020-12-22
CN112116104B CN112116104B (en) 2024-06-18

Family

ID=73801087

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010982700.0A Active CN112116104B (en) 2020-09-17 2020-09-17 Method, device, medium and electronic equipment for automatically integrating machine learning

Country Status (1)

Country Link
CN (1) CN112116104B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113837323A (en) * 2021-11-08 2021-12-24 中国联合网络通信集团有限公司 Satisfaction prediction model training method and device, electronic equipment and storage medium
CN114003393A (en) * 2021-12-30 2022-02-01 南京大学 Method and system for improving integrated automatic machine learning operation performance
WO2022193408A1 (en) * 2021-03-17 2022-09-22 中奥智能工业研究院(南京)有限公司 Automatic data analysis and modeling process

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110210556A (en) * 2019-05-29 2019-09-06 中国科学技术大学 Pedestrian identifies data creation method again
CN110222848A (en) * 2019-05-05 2019-09-10 阿里巴巴集团控股有限公司 The determination method and device for the integrated model that computer executes
CN110503208A (en) * 2019-08-26 2019-11-26 第四范式(北京)技术有限公司 Resource regulating method and resource scheduling device in multi-model exploration
CN110517492A (en) * 2019-08-27 2019-11-29 中国科学院自动化研究所 Based on the traffic route recommended method of parallel integrated study, system, device
CN110763660A (en) * 2019-10-22 2020-02-07 华南理工大学 LIBS quantitative analysis method based on ensemble learning
CN110914839A (en) * 2017-06-26 2020-03-24 D5Ai有限责任公司 Selective training of error decorrelation
US20200193313A1 (en) * 2018-12-14 2020-06-18 Parallel Machines, Inc. Interpretability-based machine learning adjustment during production
US20200242400A1 (en) * 2019-01-25 2020-07-30 Oath Inc. Systems and methods for hyper parameter optimization for improved machine learning ensembles
CN111597758A (en) * 2020-05-14 2020-08-28 河海大学 Medium and small river integrated forecasting method based on negative correlation learning

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110914839A (en) * 2017-06-26 2020-03-24 D5Ai有限责任公司 Selective training of error decorrelation
US20200193313A1 (en) * 2018-12-14 2020-06-18 Parallel Machines, Inc. Interpretability-based machine learning adjustment during production
US20200242400A1 (en) * 2019-01-25 2020-07-30 Oath Inc. Systems and methods for hyper parameter optimization for improved machine learning ensembles
CN110222848A (en) * 2019-05-05 2019-09-10 阿里巴巴集团控股有限公司 The determination method and device for the integrated model that computer executes
CN110210556A (en) * 2019-05-29 2019-09-06 中国科学技术大学 Pedestrian identifies data creation method again
CN110503208A (en) * 2019-08-26 2019-11-26 第四范式(北京)技术有限公司 Resource regulating method and resource scheduling device in multi-model exploration
CN110517492A (en) * 2019-08-27 2019-11-29 中国科学院自动化研究所 Based on the traffic route recommended method of parallel integrated study, system, device
CN110763660A (en) * 2019-10-22 2020-02-07 华南理工大学 LIBS quantitative analysis method based on ensemble learning
CN111597758A (en) * 2020-05-14 2020-08-28 河海大学 Medium and small river integrated forecasting method based on negative correlation learning

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
杨峻山;纪震;谢维信;朱泽轩;: "基于粒子群优化的生物组学数据分类模型选择", 深圳大学学报(理工版), no. 03, 30 May 2016 (2016-05-30) *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022193408A1 (en) * 2021-03-17 2022-09-22 中奥智能工业研究院(南京)有限公司 Automatic data analysis and modeling process
CN113837323A (en) * 2021-11-08 2021-12-24 中国联合网络通信集团有限公司 Satisfaction prediction model training method and device, electronic equipment and storage medium
CN113837323B (en) * 2021-11-08 2023-07-04 中国联合网络通信集团有限公司 Training method and device of satisfaction prediction model, electronic equipment and storage medium
CN114003393A (en) * 2021-12-30 2022-02-01 南京大学 Method and system for improving integrated automatic machine learning operation performance
CN114003393B (en) * 2021-12-30 2022-06-14 南京大学 Method and system for improving integrated automatic machine learning operation performance

Also Published As

Publication number Publication date
CN112116104B (en) 2024-06-18

Similar Documents

Publication Publication Date Title
CN110807515B (en) Model generation method and device
CN110766142A (en) Model generation method and device
WO2021136365A1 (en) Application development method and apparatus based on machine learning model, and electronic device
Hernández-Lobato et al. Probabilistic backpropagation for scalable learning of bayesian neural networks
CN112699991A (en) Method, electronic device, and computer-readable medium for accelerating information processing for neural network training
CN112116104B (en) Method, device, medium and electronic equipment for automatically integrating machine learning
CN111406267A (en) Neural architecture search using performance-predictive neural networks
CN110503192A (en) The effective neural framework of resource
WO2019155064A1 (en) Data compression using jointly trained encoder, decoder, and prior neural networks
CN109906460A (en) Dynamic cooperation attention network for question and answer
CN111602148A (en) Regularized neural network architecture search
US11989657B2 (en) Automated machine learning pipeline for timeseries datasets utilizing point-based algorithms
US20220188645A1 (en) Using generative adversarial networks to construct realistic counterfactual explanations for machine learning models
US11900263B2 (en) Augmenting neural networks
US11861469B2 (en) Code generation for Auto-AI
US20220198277A1 (en) Post-hoc explanation of machine learning models using generative adversarial networks
CN111373406B (en) Accelerated simulation setup process using a priori knowledge extraction of problem matching
US20190228297A1 (en) Artificial Intelligence Modelling Engine
US20230267307A1 (en) Systems and Methods for Generation of Machine-Learned Multitask Models
CN113557534A (en) Deep forest model development and training
CN112200296A (en) Network model quantification method and device, storage medium and electronic equipment
CN112420125A (en) Molecular attribute prediction method and device, intelligent equipment and terminal
CN110782016A (en) Method and apparatus for optimizing neural network architecture search
Solazzo et al. Hardware design automation of convolutional neural networks
CN114462582A (en) Data processing method, device and equipment based on convolutional neural network model

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information

Address after: Room 221, 2 / F, block C, 18 Kechuang 11th Street, Daxing District, Beijing, 100176

Applicant after: Jingdong Technology Holding Co.,Ltd.

Address before: Room 221, 2 / F, block C, 18 Kechuang 11th Street, Daxing District, Beijing, 100176

Applicant before: Jingdong Digital Technology Holding Co.,Ltd.

CB02 Change of applicant information
GR01 Patent grant
GR01 Patent grant