CN110633805A - Longitudinal federated learning system optimization method, device, equipment and readable storage medium - Google Patents

Longitudinal federated learning system optimization method, device, equipment and readable storage medium Download PDF

Info

Publication number
CN110633805A
CN110633805A CN201910918262.9A CN201910918262A CN110633805A CN 110633805 A CN110633805 A CN 110633805A CN 201910918262 A CN201910918262 A CN 201910918262A CN 110633805 A CN110633805 A CN 110633805A
Authority
CN
China
Prior art keywords
data
model
sample data
gradient information
output
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910918262.9A
Other languages
Chinese (zh)
Other versions
CN110633805B (en
Inventor
程勇
刘洋
陈天健
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
WeBank Co Ltd
Original Assignee
WeBank Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by WeBank Co Ltd filed Critical WeBank Co Ltd
Priority to CN201910918262.9A priority Critical patent/CN110633805B/en
Publication of CN110633805A publication Critical patent/CN110633805A/en
Priority to PCT/CN2020/107332 priority patent/WO2021004551A1/en
Application granted granted Critical
Publication of CN110633805B publication Critical patent/CN110633805B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising
    • G06Q30/0201Market modelling; Market analysis; Collecting market data
    • G06Q30/0202Market predictions or forecasting for commercial activities

Landscapes

  • Engineering & Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Strategic Management (AREA)
  • Development Economics (AREA)
  • Theoretical Computer Science (AREA)
  • Finance (AREA)
  • Accounting & Taxation (AREA)
  • Data Mining & Analysis (AREA)
  • Software Systems (AREA)
  • Physics & Mathematics (AREA)
  • Entrepreneurship & Innovation (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Mathematical Physics (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Medical Informatics (AREA)
  • Game Theory and Decision Science (AREA)
  • Economics (AREA)
  • Marketing (AREA)
  • General Business, Economics & Management (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Machine Translation (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a method, a device, equipment and a readable storage medium for optimizing a longitudinal federated learning system, wherein the method comprises the following steps: the method comprises the steps that sample alignment is carried out on first equipment and second equipment to obtain first sample data of the first equipment, wherein the data characteristics of the first sample data are different from those of second sample data, and the second sample data are obtained by carrying out sample alignment on the second equipment and the first equipment; and performing cooperative training by adopting the first sample data and the second equipment to obtain an interpolation model, wherein the interpolation model is used for inputting data belonging to the data characteristics corresponding to the first equipment and outputting prediction data belonging to the data characteristics corresponding to the second equipment. When the longitudinal federal learning trained model is used by the participants of longitudinal federal learning, the model can be independently used without the cooperation of other participants, and the application range of the longitudinal federal learning is expanded.

Description

Longitudinal federated learning system optimization method, device, equipment and readable storage medium
Technical Field
The invention relates to the technical field of machine learning, in particular to a method, a device, equipment and a readable storage medium for optimizing a longitudinal federated learning system.
Background
With the development of artificial intelligence, people provide a concept of 'federal learning' for solving the problem of data islanding, so that both federal parties can train a model to obtain model parameters without providing own data, and the problem of data privacy disclosure can be avoided.
In the longitudinal federated learning, under the condition that the data features of the participants are overlapped less and the users are overlapped more, the part of the users and the data with the same users and different user data features of the participants are taken out to jointly train the machine learning model. For example, there are two participants a and B belonging to the same region, where participant a is a bank and participant B is an e-commerce platform. Participants a and B have more users in the same area, but a and B have different services and different recorded user data characteristics. In particular, the user data characteristics of the a and B records may be complementary. In such a scenario, vertical federated learning may be used to help a and B build a joint machine learning predictive model, helping a and B provide better service to their customers.
However, when using a model trained through longitudinal federal learning, participant a needs to collaborate with participant B to use the model for prediction. For example, participant a has data characteristics X3, X4 and X5 and participant B has data characteristics X1 and X2, and when participant a needs to predict a new client, participant a needs to communicate with participant B to query whether participant B also has the client because participant a does not have all of the data characteristics, i.e., no data for the client under data characteristics X1 and X2, and participant a cannot predict the client if participant B does not have the client's data. Even though participant B has the client's data, participant a and B need to cooperate to make predictions about the client.
Disclosure of Invention
The invention mainly aims to provide a method, a device, equipment and a readable storage medium for optimizing a longitudinal federal learning system, aiming at realizing that a participant in longitudinal federal learning can independently use a model without the cooperation of other participants when using the model trained through longitudinal federal learning.
In order to achieve the above object, the present invention provides a longitudinal federal learning system optimization method, which is applied to a first device, wherein the first device is in communication connection with a second device, and the longitudinal federal learning system optimization method includes the following steps:
performing sample alignment with the second device to obtain first sample data of the first device, wherein the first sample data has different data characteristics from second sample data obtained by performing sample alignment between the second device and the first device;
and obtaining an interpolation model by adopting the first sample data and the second equipment for cooperative training, wherein the interpolation model is used for inputting data belonging to the data characteristics corresponding to the first equipment and outputting data belonging to the data characteristics corresponding to the second equipment.
Optionally, the step of obtaining an interpolation model by using the first sample data and the second device for training in cooperation includes:
inputting the first sample data into a first partial model preset in the first equipment to obtain a first output;
sending the first output to the second equipment, so that the second equipment obtains a second output of a preset second partial model according to the first output, calculates a first loss function and first gradient information according to the second sample data and the second output, and updates parameters of the second partial model according to gradient information related to the second partial model in the first gradient information;
updating parameters of the first part of models according to gradient information related to the first part of models in the first gradient information received from the second equipment, and iteratively training until a preset stopping condition is met, and receiving the second part of models sent by the second equipment;
and combining the first partial model and the second partial model to obtain the interpolation model.
Optionally, after the step of combining the first partial model and the second partial model to obtain the interpolation model, the method further includes:
inputting local sample data belonging to the data characteristics corresponding to the first equipment into the interpolation model to obtain prediction sample data belonging to the data characteristics corresponding to the second equipment;
and locally training a preset machine learning model to be trained by adopting the local sample data and the prediction sample data to obtain a target machine learning model.
Optionally, a first Trusted Execution Environment (TEE) module is included in the first device, a second TEE module is included in the second device,
the sending the first output to the second device, so that the second device obtains a second output of a preset second partial model according to the first output, calculates a first loss function and first gradient information according to the second sample data and the second output, and updates parameters of the second partial model according to gradient information related to the second partial model in the first gradient information, where the parameters include:
encrypting the first output to obtain a first encrypted output;
sending the first encrypted output to the second device, so that the second device decrypts the first encrypted output in the second TEE module to obtain the first output, obtains a second output preset with a second partial model according to the first output, calculates a first loss function and first gradient information according to the second sample data and the second output, updates parameters of the second partial model according to gradient information related to the second partial model in the first gradient information, and encrypts gradient information related to the first partial model in the first gradient information to obtain encrypted gradient information;
the step of updating the parameters of the first partial model according to the gradient information related to the first partial model in the first gradient information received from the second device comprises:
and receiving the encrypted gradient information sent by the second device, decrypting the encrypted gradient information in the first TEE module to obtain gradient information related to the first part model in the first gradient information, and updating parameters of the first part model according to the gradient information related to the second part model.
Optionally, after the step of sending the first output to the second device, the method further includes:
receiving the second output and the first loss function sent by the second device;
inputting the first sample data and the second output into a preset machine learning model to be trained to obtain predicted label data;
calculating a second loss function and second gradient information of the machine learning model to be trained according to the predicted label data and pre-stored local actual label data;
the step of receiving the second partial model sent by the second device when the iterative training is detected to meet the preset stop condition comprises:
and updating parameters of the machine learning model to be trained according to the second gradient information, performing iterative training to minimize a fusion loss function until a target machine learning model is obtained when a preset stopping condition is met, and receiving the second part of model sent by the second equipment, wherein the first equipment fuses the first loss function and the second loss function to obtain the fusion loss function.
Optionally, the first device includes a TEE module, and the step of obtaining an interpolation model by using the first sample data and the second device for training in cooperation includes:
receiving second encryption sample data sent by the second device, wherein the second device encrypts the second sample data to obtain the second encryption sample data;
and decrypting the second encrypted sample data in the TEE module to obtain second sample data, and training the interpolation model to be trained according to the first sample data and the second sample data to obtain the interpolation model.
Optionally, the target machine learning model is configured to predict a purchase intention of a user, and after the step of performing local training on a preset machine learning model to be trained by using the local sample data and the predicted sample data to obtain the target machine learning model, the method further includes:
inputting first data of a target user into the interpolation model to obtain second data, wherein the data characteristics of the first data comprise user identity characteristics, and the data characteristics of the second data comprise user purchase characteristics;
and inputting the first data and the second data into the target machine learning model to obtain the purchase intention of the target user.
In order to achieve the above object, the present invention further provides a longitudinal federal learning system optimization device, where the longitudinal federal learning system optimization device is disposed on a first device, and the first device is in communication connection with a second device, and the longitudinal federal learning system optimization device includes:
an alignment module, configured to perform sample alignment with the second device to obtain first sample data of the first device, where data characteristics of the first sample data are different from those of second sample data, and the second sample data is obtained by performing sample alignment between the second device and the first device;
and the training module is used for obtaining an interpolation model by adopting the first sample data and the second equipment for cooperative training, wherein the interpolation model is used for inputting data belonging to the data characteristics corresponding to the first equipment and outputting prediction data belonging to the data characteristics corresponding to the second equipment.
In order to achieve the above object, the present invention further provides a longitudinal federal learning system optimization device, including: a memory, a processor, and a longitudinal federated learning system optimization program stored on the memory and executable on the processor, the longitudinal federated learning system optimization program when executed by the processor implementing the steps of the longitudinal federated learning system optimization method as described above.
In addition, to achieve the above object, the present invention further provides a computer readable storage medium, on which a longitudinal federal learning system optimization program is stored, and the longitudinal federal learning system optimization program, when executed by a processor, implements the steps of the longitudinal federal learning system optimization method as described above.
In the invention, the first device performs sample alignment with the second device, and performs collaborative training by using the aligned sample data to obtain an interpolation model capable of complementing the missing data characteristics of the first device relative to the second device, so that when the first device uses the machine learning model obtained by training the data characteristics of the first device and the second device, even if the first device does not have the data of the second device, the first device can independently predict the data belonging to the corresponding data characteristics of the second device through the interpolation model, thereby completing the prediction by using the machine learning model through the complemented data, expanding the application range of longitudinal federal learning, and avoiding that the first device cannot use the machine learning model obtained by longitudinal federal learning under the scene that the second device cannot provide data for the first device.
Drawings
FIG. 1 is a schematic diagram of a hardware operating environment according to an embodiment of the present invention;
FIG. 2 is a schematic flow chart of a first embodiment of a longitudinal federated learning system optimization method of the present invention;
FIG. 3 is a schematic diagram of sample data according to an embodiment of the present invention;
FIG. 4 is a schematic diagram of a cutting method of a mold according to an embodiment of the present invention;
FIG. 5 is a diagram illustrating a first device alone training a machine learning model according to an embodiment of the present invention;
fig. 6 is a schematic flowchart of a process of training an interpolation model by cooperation of a first device and a second device and training a machine learning model by the first device alone according to an embodiment of the present invention;
fig. 7 is a schematic diagram illustrating a first device and a second device cooperatively training an interpolation model in a respective TEE module environment according to an embodiment of the present invention;
fig. 8 is a schematic diagram illustrating a first device and a second device cooperatively training an interpolation model and a machine learning model according to an embodiment of the present invention;
FIG. 9 is a block diagram of a preferred embodiment of the longitudinal federated learning system optimization apparatus of the present invention.
The implementation, functional features and advantages of the objects of the present invention will be further explained with reference to the accompanying drawings.
Detailed Description
It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
As shown in fig. 1, fig. 1 is a schematic device structure diagram of a hardware operating environment according to an embodiment of the present invention.
It should be noted that, in the embodiment of the present invention, the longitudinal federal learning system optimization device may be a smart phone, a personal computer, a server, and other devices, which are not specifically limited herein.
As shown in fig. 1, the longitudinal federal learning system optimization device may include: a processor 1001, such as a CPU, a network interface 1004, a user interface 1003, a memory 1005, a communication bus 1002. Wherein a communication bus 1002 is used to enable connective communication between these components. The user interface 1003 may include a Display screen (Display), an input unit such as a Keyboard (Keyboard), and the optional user interface 1003 may also include a standard wired interface, a wireless interface. The network interface 1004 may optionally include a standard wired interface, a wireless interface (e.g., WI-FI interface). The memory 1005 may be a high-speed RAM memory or a non-volatile memory (e.g., a magnetic disk memory). The memory 1005 may alternatively be a storage device separate from the processor 1001.
Those skilled in the art will appreciate that the configuration of the apparatus shown in FIG. 1 does not constitute a limitation on the longitudinal Federal learning System optimization apparatus, and may include more or fewer components than shown, or some components in combination, or a different arrangement of components.
As shown in fig. 1, the memory 1005, which is a type of computer storage medium, may include an operating system, a network communication module, a user interface module, and a longitudinal federal learning system optimization program therein. A TEE (Trusted execution environment) module is also included. The operating system is a program for managing and controlling hardware and software resources of the equipment and supports the running of a federally learned private data processing program and other software or programs. The TEE is a secure area within the host processor, running in a separate environment and running in parallel with the operating system, which ensures that the confidentiality and integrity of the code and data loaded in the TEE are protected. Trusted applications running in the TEE can access all functions of the device main processor and memory, while hardware isolation protects these components from user-installed applications running in the main operating system. In this embodiment, the TEE module may be implemented in various ways, such as Software guard extensions (SGX) based on Intel, Secure Encrypted Virtualization (SEV) of AMD, Trust Zone of ARM, or santtum of MIT. Authentication and authorization of the TEE module may be accomplished through a third party secure server. For example, when a TEE is an SGX using Intel, the TEE may be authenticated by the security server of Intel, i.e., the TEE is secured.
In the device shown in fig. 1, the user interface 1003 is mainly used for data communication with a client; the network interface 1004 is mainly used for establishing communication connection with other terminal devices participating in federal learning, such as a second device which is a participant in longitudinal federal learning; and the processor 1001 may be configured to invoke a longitudinal federated learning system optimization program stored in the memory 1005 and perform the following operations:
performing sample alignment with the second device to obtain first sample data of the first device, wherein the first sample data has different data characteristics from second sample data obtained by performing sample alignment between the second device and the first device;
and obtaining an interpolation model by adopting the first sample data and the second equipment for cooperative training, wherein the interpolation model is used for inputting data belonging to the data characteristics corresponding to the first equipment and outputting data belonging to the data characteristics corresponding to the second equipment.
Further, the step of obtaining an interpolation model by using the first sample data and the second device for training in cooperation includes:
inputting the first sample data into a first partial model preset in the first equipment to obtain a first output;
sending the first output to the second equipment, so that the second equipment obtains a second output of a preset second partial model according to the first output, calculates a first loss function and first gradient information according to the second sample data and the second output, and updates parameters of the second partial model according to gradient information related to the second partial model in the first gradient information;
updating parameters of the first part of models according to gradient information related to the first part of models in the first gradient information received from the second equipment, and iteratively training until a preset stopping condition is met, and receiving the second part of models sent by the second equipment;
and combining the first partial model and the second partial model to obtain the interpolation model.
Further, after the step of combining the first partial model and the second partial model to obtain the interpolation model, the processor 1001 may be further configured to call a federal learned privacy data processing program stored in the memory 1005, and perform the following steps:
inputting local sample data belonging to the data characteristics corresponding to the first equipment into the interpolation model to obtain prediction sample data belonging to the data characteristics corresponding to the second equipment;
and locally training a preset machine learning model to be trained by adopting the local sample data and the prediction sample data to obtain a target machine learning model.
Further, a first TEE module is included in the first device, a second TEE module is included in the second device,
the sending the first output to the second device, so that the second device obtains a second output of a preset second partial model according to the first output, calculates a first loss function and first gradient information according to the second sample data and the second output, and updates parameters of the second partial model according to gradient information related to the second partial model in the first gradient information, where the parameters include:
encrypting the first output to obtain a first encrypted output;
sending the first encrypted output to the second device, so that the second device decrypts the first encrypted output in the second TEE module to obtain the first output, obtains a second output preset with a second partial model according to the first output, calculates a first loss function and first gradient information according to the second sample data and the second output, updates parameters of the second partial model according to gradient information related to the second partial model in the first gradient information, and encrypts gradient information related to the first partial model in the first gradient information to obtain encrypted gradient information;
the step of updating the parameters of the first partial model according to the gradient information related to the first partial model in the first gradient information received from the second device comprises:
and receiving the encrypted gradient information sent by the second device, decrypting the encrypted gradient information in the first TEE module to obtain gradient information related to the first part model in the first gradient information, and updating parameters of the first part model according to the gradient information related to the first part model.
Further, after the step of sending the first output to the second device, the processor 1001 may be further configured to call a federally learned privacy data handler stored in the memory 1005, and perform the following steps:
receiving the second output and the first loss function sent by the second device;
inputting the first sample data and the second output into a preset machine learning model to be trained to obtain predicted label data;
calculating a second loss function and second gradient information of the machine learning model to be trained according to the predicted label data and pre-stored local actual label data;
the step of receiving the second partial model sent by the second device when the iterative training is detected to meet the preset stop condition comprises:
and updating parameters of the machine learning model to be trained according to the second gradient information, performing iterative training to minimize a fusion loss function until a target machine learning model is obtained when a preset stopping condition is met, and receiving the second part of model sent by the second equipment, wherein the first equipment fuses the first loss function and the second loss function to obtain the fusion loss function.
Further, the first device includes a TEE module, and the step of obtaining an interpolation model by using the first sample data and the second device for training in cooperation includes:
receiving second encryption sample data sent by the second device, wherein the second device encrypts the second sample data to obtain the second encryption sample data;
and decrypting the second encrypted sample data in the TEE module to obtain second sample data, and training the interpolation model to be trained according to the first sample data and the second sample data to obtain the interpolation model.
Further, the target machine learning model is configured to predict a purchase intention of a user, and after the step of performing local training on a preset machine learning model to be trained by using the local sample data and the prediction sample data to obtain the target machine learning model, the processor 1001 may be further configured to call a federal learning privacy data processing program stored in the memory 1005, and execute the following steps:
inputting first data of a target user into the interpolation model to obtain second data, wherein the data characteristics of the first data comprise user identity characteristics, and the data characteristics of the second data comprise user purchase characteristics;
and inputting the first data and the second data into the target machine learning model to obtain the purchase intention of the target user.
Based on the structure, various embodiments of the longitudinal federal learning system optimization method are provided.
While a logical order is shown in the flow chart, in some cases, the steps shown or described may be performed in an order different than presented herein. The first device and the second device related in the embodiment of the present invention may be participating devices participating in federal learning of longitudinal federal learning, and the participating devices may be devices such as a smart phone, a personal computer, and a server.
Referring to fig. 2, fig. 2 is a schematic flow chart of a first embodiment of the longitudinal federated learning system optimization method of the present invention. In this embodiment, the longitudinal federated learning system optimization method includes:
step S10, performing sample alignment with the second device to obtain first sample data of the first device, where the first sample data and second sample data have different data characteristics, and the second sample data is obtained by performing sample alignment between the second device and the first device;
it should be noted that, in the following description, the first device and the second device distinguish two participating devices, and a model required for training and using the first device is taken as an example. The model required by the second device can be trained and used only by exchanging the roles of the first device and the second device, and the method is applicable to a scene with a plurality of participating devices and coordination devices and participating devices according to similar principles.
The first device establishes a communication connection with the second device in advance. The local data of the first device and the second device have an overlapping part in user dimension, and have different parts (possibly completely different) in data characteristics, the first device and the second device perform sample alignment by adopting respective local data to determine common users and different data characteristics of the two devices, the first device takes the data of the common users in the local data as first sample data, and the second device takes the data of the common users in the local data, which is different from the data characteristics of the first device, as second sample data, that is, the finally determined first sample data and the second sample data have the same user and different data characteristics. The first device and the second device may perform sample alignment by using an existing sample alignment technique, and if the first device and the second device are not trusted, an encrypted sample alignment technique may be used, which is not described in detail herein. For example, fig. 3 is a schematic diagram of sample data in a first device and a second device, where the local data of the first device includes 3 users { U1, U2, U3}, the data characteristics include { X1, X2, X3}, the local data of the second device includes 3 users { U1, U2, U4}, and the data characteristics include { X4, X5 }. After sample alignment, the first sample data determined by the first device is data of users U1 and U2 under data characteristics X1, X2 and X3, and the second sample data determined by the second device is data of users U1 and U2 under data characteristics X4 and X5.
Step S20, obtaining an interpolation model by using the first sample data and the second device for collaborative training, where the interpolation model is used to input data belonging to the data feature corresponding to the first device and output data belonging to the data feature corresponding to the second device.
The first device adopts the first sample data to cooperatively train with the second device to obtain an interpolation (interpolation) model. The interpolation model is used for complementing data of the missing data characteristics of the first device relative to the second device. In the above specific example, the interpolation model is used to complement the data of the first sample data of the first device under the data characteristics X4 and X5 of the users U1 and U2. Specifically, the interpolation model may be configured to input data belonging to a data feature corresponding to the first device and output data belonging to a data feature corresponding to the second device, that is, data of the user in the first sample data is input into the interpolation model, and data close to the user in the second sample data is output, for example, data of the user U1 under data features X1, X2, and X3 is input, and the interpolation model outputs data close to data of the user U1 under data features X4 and X5 in the second sample data, so as to complement data of the missing user U1 under data features X4 and X5 for the first device. The interpolation model can adopt a model structure such as a variational automatic encoder (variational automatic encoder), a generation countermeasure Network (general adaptive Network), a Pixel-RNN, a Pixel-CNN or a Restricted Boltzmann Machine (Restricted Boltzmann Machine), and the like, and the first device and the second device can adopt a Split Learning (Split Learning) mode to cooperatively train the interpolation model.
Further, the step S20 includes:
step S201, inputting the first sample data into a first part of models preset in the first equipment to obtain a first output;
the method comprises the steps of presetting the structure of an interpolation model to be trained, setting the input characteristic of the interpolation model to be trained as the data characteristic corresponding to first equipment, setting the output characteristic as the data characteristic corresponding to second equipment, and cutting the interpolation model to be trained into two parts, namely a first part model and a second part model. The first partial model is pre-positioned in a first device and the second partial model is pre-positioned in a second device. Specifically, as shown in fig. 4, when the interpolation model to be trained is a neural network model including a plurality of layers, there are two ways to cut the interpolation model to be trained; the first method is to select two layers in the neural network model, cut the two layers, take the front part of the neural network model as the first part model and the back part as the second part model, and the two layers are called as cut layers; the second is to select two layers in the neural network model, add one layer between the two layers, the input of the layer is the output of the previous layer in the two layers, the output is the input of the next layer in the two layers, that is, the previous layer and the added layer are directly connected, cut the neural network model from the position between the previous layer and the added layer, the front part is used as the first part model, the back part is used as the second part model, and the position between the previous layer and the added layer is called as the cut layer; compared with the first cutting mode, the second cutting mode has less connection between the first part model and the second part model, so that the data volume needing to be transmitted between the first equipment and the second equipment is reduced, and which cutting mode is selected according to the specific situations of the dimension of the data characteristics of the first equipment and the second equipment, the structural complexity of the interpolation model to be trained and the like.
The first device inputs the first sample data into a first partial model of the first device, the first partial model outputting a first output, i.e. an output of a last layer of the first partial model.
Step S202, sending the first output to the second device, so that the second device can obtain a second output of a preset second partial model according to the first output, calculate a first loss function and first gradient information according to the second sample data and the second output, and update parameters of the second partial model according to gradient information related to the second partial model in the first gradient information;
the first device sends the first output to the second device. It should be noted that, if the interpolation model to be trained is cut according to the first cutting mode, the first device multiplies the first output by the corresponding connection weight according to the connection relationship between the cut layers to convert the first output into the input of the first layer of the second partial model, and sends the converted result to the second device, so that the second device inputs the converted result into the first layer of the second partial model, and the second partial model outputs the second output; and if the interpolation model to be trained is cut according to the second cutting mode, the first equipment directly sends the first output to the second equipment, the second equipment directly inputs the first output into the first layer of the second part model after receiving the first output, and the second part model outputs the second output.
And after obtaining second output of the second partial model according to the first output, the second equipment calculates a first loss function and first gradient information according to second sample data and the second output. The first loss function is a loss function of the interpolation model to be trained, and the first gradient information may include gradient information related to the second partial model, such as gradient information of the first loss function on each parameter of the second partial model, and may also include gradient information related to the first partial model, such as gradient information of the first loss function on an input variable of the second partial model.
The second device may detect whether a preset stop condition is met, where the preset stop condition may be a preset stop condition, such as stopping when the first loss function convergence is detected, or stopping when the number of iterative training times reaches a maximum number, or stopping when the time of iterative training reaches a maximum training time. Specifically, the second device may determine that the first loss function converges when detecting that the loss value of the first loss function is smaller than a preset value, and determine that the first loss function does not converge if the loss value of the first loss function is not smaller than the preset value. And when the preset stopping condition is not met, the second equipment updates the parameters in the second part of the model according to the gradient information related to the second part of the model in the first gradient information. The second device may send the gradient information related to the first partial model in the first gradient information to the first device, or may send both the first gradient information to the first device.
Step S203, updating parameters of the first part model according to gradient information related to the first part model in the first gradient information received from the second device, and receiving the second part model sent by the second device when iterative training is detected to meet a preset stop condition;
the first device receives the first gradient information from the second device, or receives the gradient information related to the first partial model in the first gradient information. And the first equipment updates each parameter of the first part model according to the gradient information related to the first part model, for example, according to the gradient information of the first loss function on the input variable of the second part model, the gradient information of the first loss function on the output variable of the first part model is deduced, the gradient information of the first loss function on each parameter of the first part model is deduced, and each parameter is updated according to the gradient information corresponding to the parameter. After the parameters of the first part of models are updated, the first device continues to input the first part of models by adopting first sample data to obtain first output, the first output is sent to the second device, the second device calculates a first loss function and gradient information, the process is circulated until the preset stopping condition is met, the training of the interpolation model to be trained is stopped, and the current parameters are used as the final parameters of the first part of models and the second part of models. The second device sends the second partial model, which determines the final parameters, to the first device. It should be noted that the second device may detect whether the preset stop condition is met, or the first device detects whether the preset stop condition is met, and if the preset stop condition is met, the second device may send the calculation result of the first loss function to the first device, and the first device detects whether the first loss function converges, or the second device sends the result of detecting whether the first loss function converges to the first device.
Step S204, combining the first partial model and the second partial model to obtain the interpolation model.
And the first equipment receives the second partial model which is sent by the second equipment and determines the final parameters, and combines the second partial model with the first partial model determining the final parameters to obtain the trained interpolation model. Specifically, the combination mode may correspond to a cutting mode, and when a first cutting mode is adopted, the trained first partial model and the trained second partial model are directly spliced, and when a second cutting mode is adopted, the first layer of the second partial model is removed and then spliced with the first partial model.
In this embodiment, the first device performs sample alignment with the second device, performs collaborative training by using aligned sample data, and obtains an interpolation model capable of complementing missing data features of the first device relative to the second device, so that when the first device uses a machine learning model obtained by training data features of the first device and the second device, even if there is no data of the second device, the first device can independently predict data belonging to corresponding data features of the second device through the interpolation model locally, and thus the complemented data is used to complete prediction by using the machine learning model, thereby expanding the application range of longitudinal federal learning, and avoiding that the first device cannot use the machine learning model obtained by longitudinal federal learning in a scene where the second device cannot provide data for the first device.
Further, based on the first embodiment, a second embodiment of the optimization method for a longitudinal federal learning system of the present invention is provided, and after the step S204, the second embodiment of the optimization method for a longitudinal federal learning system of the present invention further includes:
step S205, inputting local sample data belonging to the corresponding data characteristics of the first device into the interpolation model to obtain prediction sample data belonging to the corresponding data characteristics of the second device;
after the interpolation model is obtained by the first device and the second device in cooperation training, the machine learning model with the input characteristics including the data characteristics of the first device and the second device can be independently trained. Specifically, the first device locally stores local sample data (which may include data of multiple users) belonging to corresponding data features of the first device. It should be noted that the local sample data may include user data that is not used in the training interpolation model, that is, the second device may not have data of the user. And the first equipment inputs the local sample data into the interpolation model to obtain the prediction sample data belonging to the corresponding data characteristics of the second equipment. Or the first device inputs the first sample data into the interpolation model to obtain the prediction sample data which is corresponding to the second sample data and belongs to the data characteristics corresponding to the second device. For example, the first device inputs the data of the user U1 under the data features X1, X2 and X3 into the interpolation model, and obtains the data under the data features X4 and X5 (called prediction sample data), because the data are the interpolation model trained in advance, the interpolation model outputs the obtained data under the data features X4 and X5, and the data are close to the real data of the user U1 under the data features X4 and X5.
And S206, locally training a preset machine learning model to be trained by adopting the local sample data and the prediction sample data to obtain a target machine learning model.
The first equipment carries out local training on a preset machine learning model to be trained by adopting local sample data and prediction sample data to obtain a target machine learning model. Or the first device performs local training on a preset machine learning model to be trained by adopting the first sample data and the prediction sample data to obtain a target machine learning model. The machine learning model to be trained may be a machine learning model with specific prediction or classification targets, such as a neural network model. Specifically, the first device combines the data features of the user data in the local sample data and the data features of the user data in the prediction sample data to serve as input features of the machine learning model to be trained, iterative training is performed on the machine learning model to be trained until the machine learning model to be trained is detected to be converged, and the target machine learning model is obtained after training is completed. The first device may use the trained target machine learning model to perform a prediction or classification task.
It should be noted that, if the machine learning model to be trained is a supervised machine learning model and only the second device has a data tag, the second device may share the data corresponding to the data tag with the first device, and the first device performs supervised training on the machine learning model to be trained by using the data corresponding to the data tag. If the first device and the second device are in an untrusted scene, the second device may share the data tag with the first device in an encrypted manner.
Fig. 5 is a schematic diagram of the first device training a machine learning model by itself. GM in the figureAIs an interpolation model obtained by the cooperative training of a first device and a second device, wherein the first device uses the first sample data XAInput GMAAnd MALocally learning the model M to the machine by means of an interpolation modelATraining is carried out, and the first sample data X isAInput GMAAnd MA
In this embodiment, the first device completes missing data features of the first device by using an interpolation model obtained through collaborative training with the second device, so that the first device can independently train locally to obtain a target machine learning model even without the help of the second device, and thus, the application scenario of longitudinal federal learning is expanded.
Further, the target machine learning model is used for predicting the purchase intention of the user, and after the step S206, the method further includes:
step S207, inputting first data of a target user into the interpolation model to obtain second data, wherein the data characteristics of the first data comprise user identity characteristics, and the data characteristics of the second data comprise user purchase characteristics;
the target machine learning model may be a machine learning model for predicting the purchasing intention of the user, that is, the output label of the target machine learning model may be the purchasing intention, such as an output result of 0 or 1, where 1 indicates that the user will purchase and 0 indicates that the user will not purchase. The first device may be a device deployed with a banking institution, the second device may be a device deployed with an e-commerce institution, and the data characteristics of the user data in the first device may be different from those in the second device due to different businesses, and the data characteristics of the user data in the first device may include user identification characteristics such as age, deposit, monthly salary and the like, and the data characteristics of the user data in the second device may include user purchase characteristics such as purchase times, purchase preferences and the like. The first device and the second device cooperate in advance to train to obtain an interpolation model for predicting user purchase characteristic data according to the user identity characteristic data.
After the target machine learning model is obtained through training, the first device can independently adopt the interpolation model and the target machine learning model to predict the purchase intention of the target user under the condition that the second device does not provide data on the aspect of the user purchase characteristics. Specifically, the first device may input first data of the target user into the interpolation model, and the interpolation model outputs second data. The first data, namely the data of the target user recorded by the first device, includes the user identity characteristic, and the second data, namely the predicted purchase characteristic data of the target user.
Step S208, inputting the first data and the second data into the target machine learning model to obtain the purchase intention of the target user.
And after the first equipment obtains the second data, inputting the first data and the second data into the target machine learning model to obtain the purchase intention of the target user. The second data is predicted by a pre-trained interpolation model and is close to the real data of the user, so that the predicted purchasing intention error of the target user is small by inputting the first data and the second data into the target machine learning model.
It should be noted that the target machine learning model may also be used in other application scenarios besides the purchasing intention prediction, such as performance level prediction, paper value evaluation, and the like, and the embodiments of the present invention are not limited herein.
Fig. 6 is a schematic flow chart of a process of training an interpolation model by cooperation of a first device and a second device, and training a machine learning model by the first device alone.
In this embodiment, an interpolation model is obtained through cooperative training of the first device and the second device, and when a target machine learning model is used, the interpolation model is used to complement the missing data characteristics of the first device, so that the first device can use the target machine learning model to complete a prediction function even in a scene without cooperation of the second device, and thus the application range of longitudinal federal learning is expanded.
Further, based on the first and second embodiments, a third embodiment of the optimization method for a longitudinal federated learning system according to the present invention is provided, where in the third embodiment of the optimization method for a longitudinal federated learning system according to the present invention, the first device includes a first TEE module, the second device includes a second TEE module, and the step S202 includes:
step S2021, encrypting the first output to obtain a first encrypted output;
in this embodiment, to adapt to a scenario in which the first device and the second device are not trusted with each other, the first device and the second device may be respectively provided with a TEE module, and the private data is processed in the TEE module, and the private data outside the TEE module is encrypted, so that the private data of the other party cannot be acquired. Specifically, after obtaining the first output, the first device may encrypt the first output to obtain a first encrypted output, and the encryption manner is not limited in this embodiment. It should be noted that, because the first output is obtained according to the local data of the first device and is visible in the first device, the first output may be encrypted in the first TEE module or encrypted in the second TEE module, and the encryption is to ensure that the second device cannot obtain the original first output after sending the first output to the second device, so that it is ensured that the data information of the first device is not leaked to the second device in a scenario where the first device and the second device are not trusted.
Step S2022, sending the first encrypted output to the second device, so that the second device decrypts the first encrypted output in the second TEE module to obtain the first output, obtains a second output preset with a second partial model according to the first output, calculates a first loss function and first gradient information according to the second sample data and the second output, updates parameters of the second partial model according to gradient information related to the second partial model in the first gradient information, and encrypts gradient information related to the first partial model in the first gradient information to obtain encrypted gradient information;
the first device outputs the first encryption to the second device. And after receiving the first encrypted output, the second equipment decrypts the first encrypted output in the second TEE module to obtain a first output. It should be noted that the decryption mode in the second TEE module of the second device corresponds to the encryption mode in the first TEE module of the first device, and if the first output is encrypted by using the public key in the first TEE module, the second TEE module decrypts by using the private key corresponding to the public key. After the second device decrypts the first output in the second TEE module, because the first output cannot be exposed to the part outside the second TEE module, the second device continues to obtain the first loss function and the first gradient information in the TEE module in a manner similar to that in step S202, and updates the parameters of the second partial model, and encrypts the gradient information related to the first partial model in the first gradient information to obtain encrypted gradient information. And after obtaining the encryption gradient information, the second equipment sends the encryption gradient information to the first equipment. Because the first output, the first loss function and the first gradient information are obtained on the basis of data in the first device, the original first output, the first loss function and the first gradient information are processed in the TEE module in the second device, and the data are encrypted outside the TEE module of the second device, so that the second device cannot obtain the private data of the first device.
The step of updating the parameters of the first partial model according to the gradient information related to the first partial model in the first gradient information received from the second device in the step SS203 includes:
step S2031, receiving the encrypted gradient information sent by the second device, decrypting the encrypted gradient information in the first TEE module to obtain gradient information related to the first partial model in the first gradient information, and updating parameters of the first partial model according to the gradient information related to the first partial model.
After the first device receives the encrypted gradient information sent by the second device, because the encrypted gradient information is obtained on the basis of the data of the second device, in order to ensure that the data of the second device is not leaked to the first device, the first device decrypts the encrypted gradient information in the first TEE module, and obtains the gradient information related to the first part model in the original first gradient information. The way of decrypting the data in the first TEE module corresponds to the way of encrypting the data in the second TEE module. After obtaining the gradient information related to the first partial model in the first gradient information, the first device updates the parameters of the first partial model according to the gradient information related to the first partial model.
After updating the parameters of the first part of models, the first equipment continues to input the first part of models by adopting first sample data to obtain first output, the first output is encrypted and sent to the second equipment, the second equipment calculates a first loss function and gradient information in the TEE module, the process is circulated until the preset stopping condition is met, the training of the interpolation model to be trained is stopped, and the current parameters are used as the final parameters of the first part of models and the second part of models. The second device sends a second partial model (encryptable) determining the final parameters to the first device. And the first equipment obtains an interpolation model according to the combination of the first partial model and the second partial model. Specifically, similar to the process of step S204, detailed description is omitted here.
FIG. 6 is a schematic diagram of a first device and a second device cooperatively training an interpolation model in a respective TEE module environment, where the first device uses a first sample data X in a first TEE moduleAInputting interpolation model GMAThe first partial model GM ofA-in Part1, the first output is encrypted and then transmitted to the second device, and the second device decrypts the first output in the second TEE module and obtains the second Part model GM according to the decryption resultASecond output X of Part2B' and calculating gradient information, encrypting the gradient information, then transmitting the gradient information to the first equipment, updating parameters by the first equipment according to the gradient information, and carrying out iterative training until meeting a preset stop condition, wherein the GM is processed by the second equipmentA-Part2 to the first device, the first device sending a GMA-Part1 and GMA-Part2 combination to obtain a trained interpolation model GMA
In this embodiment, the TEE modules are respectively arranged in the first device and the second device, the first device and the second device process original data in the TEE modules, and only encrypted data can be obtained outside the TEE modules, so that private data cannot be leaked to the other side in a mutually untrusted scene of the first device and the second device, and the safety of the data is guaranteed.
Further, based on the first, second, and third embodiments, a fourth embodiment of the optimization method for a longitudinal federated learning system according to the present invention is provided, where in the fourth embodiment of the optimization method for a longitudinal federated learning system according to the present invention, after the step S202, the method further includes:
a step a10 of receiving the second output and the first loss function sent by the second device;
in this embodiment, the first device and the second device may cooperatively train the machine learning model while cooperatively training the interpolation model. Specifically, on the basis of the first embodiment, after the first device sends the first output to the second device, the second device obtains the second output, and calculates the first loss function and the first gradient information, the second device sends the second output, the first loss function, and the second gradient information to the first device.
Step A20, inputting the first sample data and the second output into a preset machine learning model to be trained to obtain predicted label data;
the first equipment is provided with a machine learning model to be trained in advance. And the first equipment inputs the first sample data and the second output into the machine learning model to be trained, and the machine learning model to be trained outputs the prediction label data.
Step A30, calculating a second loss function and second gradient information of the machine learning model to be trained according to the predicted label data and pre-stored local actual label data;
and the first equipment calculates a second loss function and second gradient information of the machine learning model to be trained according to the predicted label data and the pre-stored local actual label data. It should be noted that the local actual tag data pre-stored in the first device may be tag data locally recorded by the first device, or the second device may share the locally recorded tag data to the first device, where the first device is stored locally.
The step of receiving the second partial model sent by the second device until the iterative training in step S203 detects that a preset stop condition is met includes:
step S2032, updating parameters of the machine learning model to be trained according to the second gradient information, performing iterative training to minimize a fusion loss function until a target machine learning model is obtained when a preset stop condition is detected to be satisfied, and receiving the second partial model sent by the second device, wherein the first device fuses the first loss function and the second loss function to obtain the fusion loss function.
The first device fuses the first loss function and the second loss function to obtain a fusion loss function, and the fusion mode may be to calculate a weighted sum of the first loss function and the second loss function as the fusion loss function. The first device determines whether a preset stop condition is met, where the preset stop condition may be a preset stop condition, and for example, the first device stops when convergence of the fusion loss function is detected, or stops when the number of iterative training times reaches a maximum number, or stops when the time of iterative training reaches a maximum training time. Specifically, whether the fusion loss function converges may be determined according to a calculated loss value of the fusion loss function, and if the loss value of the fusion loss function is detected to be smaller than a preset value, it is determined that the fusion loss function converges, and if the loss value of the fusion loss function is not smaller than the preset value, it is determined that the fusion loss function does not converge. If the first device detects that the preset stop condition is not met, the parameters of the machine learning model to be trained can be updated according to the second gradient information, and the parameters of the first part of models can be updated according to the gradient information related to the first part of models in the first gradient information. After parameters of the first part model and the machine learning model to be trained are updated, the first equipment continues to input the first part model by adopting first sample data to obtain first output, the first output is sent to the second equipment, the second equipment calculates a first loss function and gradient information, iterative training is carried out to achieve the purpose of minimizing the fusion loss function, the training of the interpolation model to be trained and the machine learning model to be trained is stopped when the detection meets the preset stop condition, the current parameters are used as final parameters of the first part model, the second part model and the machine learning model to be trained, and the machine learning model to be trained with the finally determined parameters is used as a target machine learning model after the training is finished. And the second equipment sends the second part of model with the determined final parameters to the first equipment, and the first equipment combines the first part of model and the second part of model to obtain the trained interpolation model.
It should be noted that, on the basis of the above-mentioned collaborative training, if the first device and the second device are not trusted with each other, a TEE module may be respectively set in the first device and the second device, and the original private data is processed in the TEE module, and the private data is encrypted outside the TEE module, so as to ensure that the private data of the first device and the second device is not leaked to the other party.
As shown in fig. 7, a schematic diagram of a first device training an interpolation model and a machine learning model in cooperation with a second device is shown, and the first device uses first sample data XAInputting interpolation model GMAThe first partial model GM ofA-Part1, forwarding the first output to the second device, which derives a second partial model GM from the first outputASecond output X of Part2B', and calculates the gradient information and the loss function, forwards the second output to the first device, and backwards propagates the gradient information and the loss function to the first device. The first device updates the GM according to the gradient informationA-parameters of Part1, and taking the second output as the machine learning model M to be trainedAUntil the fusion loss function is converged, the trained machine learning model M is obtainedA. The second device sends GMA-Part2 to the first device, the first device sending a GMA-Part1 and GMA-Part2 combination to obtain a trained interpolation model GMA
In this embodiment, the machine learning model is cooperatively trained while the interpolation model is cooperatively trained by the first device and the second device, so that the training efficiency of the machine learning model is improved, and after the interpolation model is trained by the first device, missing data features can be complemented by the interpolation model, so that the first device can use the trained machine learning model to complete a prediction task under the condition that the second device does not cooperate, and the application range of longitudinal federal learning is expanded.
Further, based on the first, second, third, and fourth embodiments, a fifth embodiment of the optimization method for a longitudinal federated learning system according to the present invention is provided, where in the fifth embodiment of the optimization method for a longitudinal federated learning system according to the present invention, the first device includes a TEE module, and the step S20 includes:
step B10, receiving second encryption sample data sent by the second device, where the second device encrypts the second sample data to obtain the second encryption sample data;
in this embodiment, a different method of cooperatively training an interpolation model is proposed from that in the first embodiment. Specifically, in order to adapt to a scenario in which the first device and the second device are not trusted with each other, a TEE module may be disposed in the first device, the first device performs processing on private data in the TEE module, and the private data is encrypted outside the TEE module, so that the private data of the second device cannot be acquired. And after the second equipment performs sample alignment with the first equipment to obtain second sample data, encrypting the second sample data to obtain second encrypted data, and sending the second encrypted data to the first equipment.
And step B20, decrypting the second encrypted sample data in the TEE module to obtain second sample data, and training the interpolation model to be trained according to the first sample data and the second sample data to obtain the interpolation model.
And the first equipment decrypts the second encrypted sample data in the TEE module to obtain second sample data. It should be noted that the way of decrypting the data in the TEE module corresponds to the way of encrypting the data by the second device. The method includes that a first device trains an interpolation model to be trained in a TEE module according to first sample data and second sample data to obtain the interpolation model, and the mode of training the interpolation model by the first device alone is the same as that of training the interpolation model by the traditional device alone, and is not limited specifically here. Further, the second device may train the machine learning model locally alone in the TEE module through the trained interpolation model; the first device may also train the machine learning model while training the interpolation model in the TEE module.
In this embodiment, the TEE module is arranged in the first device, the second device encrypts second sample data and sends the second sample data to the first device, and the first device decrypts the second sample data in the TEE module to enable the second device to train the interpolation model independently under the condition that private data are not leaked to the first device, so that the situation that the interpolation model cannot be trained and used by the first device under the condition that the second device cannot cooperate with the first device to train the interpolation model is avoided, and the application range of longitudinal federal learning is expanded.
In addition, an embodiment of the present invention further provides a longitudinal federal learning system optimization device, where the longitudinal federal learning system optimization device is disposed in a first device, and the first device is in communication with a second device, and referring to fig. 9, the longitudinal federal learning system optimization device includes:
an alignment module 10, configured to perform sample alignment with the second device to obtain first sample data of the first device, where data characteristics of the first sample data are different from those of second sample data, and the second sample data is obtained by performing sample alignment between the second device and the first device;
a training module 20, configured to perform collaborative training on the first sample data and the second device to obtain an interpolation model, where the interpolation model is used to input data belonging to a data feature corresponding to the first device and output prediction data belonging to a data feature corresponding to the second device.
Further, the training module 20 includes:
the first input unit is used for inputting the first sample data into a first part of models preset in the first equipment to obtain first output;
a sending unit, configured to send the first output to the second device, so that the second device obtains a second output of a preset second partial model according to the first output, calculates a first loss function and first gradient information according to the second sample data and the second output, and updates a parameter of the second partial model according to gradient information related to the second partial model in the first gradient information;
a first receiving unit, configured to update parameters of the first partial model according to gradient information related to the first partial model in the first gradient information received from the second device, and perform iterative training until a preset stop condition is detected to be met, and receive the second partial model sent by the second device;
and the combination unit is used for combining the first partial model and the second partial model to obtain the interpolation model.
Further, the training module 20 further includes:
the second input unit is used for inputting local sample data belonging to the data characteristics corresponding to the first equipment into the interpolation model to obtain prediction sample data belonging to the data characteristics corresponding to the second equipment;
and the first training unit is used for carrying out local training on a preset machine learning model to be trained by adopting the local sample data and the prediction sample data to obtain a target machine learning model.
Further, the first device includes a first trusted execution environment TEE module therein, the second device includes a second TEE module therein, and the sending unit includes:
the encryption subunit is used for encrypting the first output to obtain a first encrypted output;
a sending subunit, configured to send the first encrypted output to the second device, so that the second device decrypts the first encrypted output in the second TEE module to obtain the first output, obtains a second output preset in a second partial model according to the first output, calculates a first loss function and first gradient information according to the second sample data and the second output, updates parameters of the second partial model according to gradient information related to the second partial model in the first gradient information, and encrypts gradient information related to the first partial model in the first gradient information to obtain encrypted gradient information;
the first receiving unit is further configured to receive the encrypted gradient information sent by the second device, decrypt the encrypted gradient information in the first TEE module to obtain gradient information in the first gradient information, where the gradient information is related to the first partial model, and update parameters of the first partial model according to the gradient information related to the first partial model.
Further, the training module 20 further includes:
a second receiving unit, configured to receive the second output and the first loss function sent by the second device;
the third input unit is used for inputting the first sample data and the second output data into a preset machine learning model to be trained to obtain predicted label data;
the calculation unit is used for calculating a second loss function and second gradient information of the machine learning model to be trained according to the predicted label data and pre-stored local actual label data;
the first receiving unit is further configured to update parameters of the machine learning model to be trained according to the second gradient information, perform iterative training to minimize a fusion loss function until a target machine learning model is obtained when a preset stop condition is detected to be met, and receive the second partial model sent by the second device, where the first device fuses the first loss function and the second loss function to obtain the fusion loss function.
Further, a TEE module is included in the first device, and the training module 20 includes:
a third receiving unit, configured to receive second encryption sample data sent by the second device, where the second device encrypts the second sample data to obtain the second encryption sample data;
and the second training unit is used for decrypting the second encrypted sample data in the TEE module to obtain second sample data, and training the interpolation model to be trained according to the first sample data and the second sample data to obtain the interpolation model.
Further, the target machine learning model is used for predicting the purchase intention of the user, and the longitudinal federal learning system optimization device further comprises:
the input module is used for inputting first data of a target user into the interpolation model to obtain second data, wherein the data characteristics of the first data comprise user identity characteristics, and the data characteristics of the second data comprise user purchase characteristics;
and the prediction module is used for inputting the first data and the second data into the target machine learning model to obtain the purchase intention of the target user.
The development content of the specific implementation mode of the longitudinal federal learning system optimization device is basically the same as that of each embodiment of the longitudinal federal learning system optimization method, and is not described herein again.
In addition, an embodiment of the present invention further provides a computer-readable storage medium, where a longitudinal federated learning system optimization program is stored on the storage medium, and when being executed by a processor, the longitudinal federated learning system optimization program implements the steps of the longitudinal federated learning system optimization method described below.
For the embodiments of the longitudinal federated learning system optimization apparatus and the computer-readable storage medium of the present invention, reference may be made to the embodiments of the longitudinal federated learning system optimization method of the present invention, which are not described herein again.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
The above-mentioned serial numbers of the embodiments of the present invention are merely for description and do not represent the merits of the embodiments.
Through the above description of the embodiments, those skilled in the art will clearly understand that the method of the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but in many cases, the former is a better implementation manner. Based on such understanding, the technical solutions of the present invention may be embodied in the form of a software product, which is stored in a storage medium (such as ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal device (such as a mobile phone, a computer, a server, an air conditioner, or a network device) to execute the method according to the embodiments of the present invention.
The above description is only a preferred embodiment of the present invention, and not intended to limit the scope of the present invention, and all modifications of equivalent structures and equivalent processes, which are made by using the contents of the present specification and the accompanying drawings, or directly or indirectly applied to other related technical fields, are included in the scope of the present invention.

Claims (10)

1. A longitudinal federated learning system optimization method is applied to a first device, the first device is in communication connection with a second device, and the longitudinal federated learning system optimization method comprises the following steps:
performing sample alignment with the second device to obtain first sample data of the first device, wherein the first sample data has different data characteristics from second sample data obtained by performing sample alignment between the second device and the first device;
and obtaining an interpolation model by adopting the first sample data and the second equipment for cooperative training, wherein the interpolation model is used for inputting data belonging to the data characteristics corresponding to the first equipment and outputting data belonging to the data characteristics corresponding to the second equipment.
2. The method of claim 1, wherein the step of training an interpolation model using the first sample data in cooperation with the second device comprises:
inputting the first sample data into a first partial model preset in the first equipment to obtain a first output;
sending the first output to the second equipment, so that the second equipment obtains a second output of a preset second partial model according to the first output, calculates a first loss function and first gradient information according to the second sample data and the second output, and updates parameters of the second partial model according to gradient information related to the second partial model in the first gradient information;
updating parameters of the first part of models according to gradient information related to the first part of models in the first gradient information received from the second equipment, and iteratively training until a preset stopping condition is met, and receiving the second part of models sent by the second equipment;
and combining the first partial model and the second partial model to obtain the interpolation model.
3. The method for optimizing a longitudinal federal learning system as claimed in claim 2, wherein said step of combining said first model portion and said second model portion to obtain said interpolation model further comprises:
inputting local sample data belonging to the data characteristics corresponding to the first equipment into the interpolation model to obtain prediction sample data belonging to the data characteristics corresponding to the second equipment;
and locally training a preset machine learning model to be trained by adopting the local sample data and the prediction sample data to obtain a target machine learning model.
4. The longitudinal federated learning system optimization method of claim 2, wherein a first Trusted Execution Environment (TEE) module is included in the first device, a second TEE module is included in the second device,
the sending the first output to the second device, so that the second device obtains a second output of a preset second partial model according to the first output, calculates a first loss function and first gradient information according to the second sample data and the second output, and updates parameters of the second partial model according to gradient information related to the second partial model in the first gradient information, where the parameters include:
encrypting the first output to obtain a first encrypted output;
sending the first encrypted output to the second device, so that the second device decrypts the first encrypted output in the second TEE module to obtain the first output, obtains a second output preset with a second partial model according to the first output, calculates a first loss function and first gradient information according to the second sample data and the second output, updates parameters of the second partial model according to gradient information related to the second partial model in the first gradient information, and encrypts gradient information related to the first partial model in the first gradient information to obtain encrypted gradient information;
the step of updating the parameters of the first partial model according to the gradient information related to the first partial model in the first gradient information received from the second device comprises:
and receiving the encrypted gradient information sent by the second device, decrypting the encrypted gradient information in the first TEE module to obtain gradient information related to the first part model in the first gradient information, and updating parameters of the first part model according to the gradient information related to the first part model.
5. The longitudinal federal learning system optimization method of claim 2, wherein said step of sending said first output to said second device is followed by further comprising:
receiving the second output and the first loss function sent by the second device;
inputting the first sample data and the second output into a preset machine learning model to be trained to obtain predicted label data;
calculating a second loss function and second gradient information of the machine learning model to be trained according to the predicted label data and pre-stored local actual label data;
the step of receiving the second partial model sent by the second device when the iterative training is detected to meet the preset stop condition comprises:
and updating parameters of the machine learning model to be trained according to the second gradient information, performing iterative training to minimize a fusion loss function until a target machine learning model is obtained when a preset stopping condition is met, and receiving the second part of model sent by the second equipment, wherein the first equipment fuses the first loss function and the second loss function to obtain the fusion loss function.
6. The method of claim 1, wherein the first device includes a TEE module, and the step of training the interpolation model using the first sample data in cooperation with the second device includes:
receiving second encryption sample data sent by the second device, wherein the second device encrypts the second sample data to obtain the second encryption sample data;
and decrypting the second encrypted sample data in the TEE module to obtain second sample data, and training the interpolation model to be trained according to the first sample data and the second sample data to obtain the interpolation model.
7. The method for optimizing a longitudinal federated learning system of claim 3, wherein the target machine learning model is used for predicting purchasing intentions of a user, and after the step of locally training a preset machine learning model to be trained using the local sample data and the prediction sample data to obtain the target machine learning model, the method further comprises:
inputting first data of a target user into the interpolation model to obtain second data, wherein the data characteristics of the first data comprise user identity characteristics, and the data characteristics of the second data comprise user purchase characteristics;
and inputting the first data and the second data into the target machine learning model to obtain the purchase intention of the target user.
8. A longitudinal federated learning system optimization apparatus deployed on a first device in communication with a second device, the longitudinal federated learning system optimization apparatus comprising:
an alignment module, configured to perform sample alignment with the second device to obtain first sample data of the first device, where data characteristics of the first sample data are different from those of second sample data, and the second sample data is obtained by performing sample alignment between the second device and the first device;
and the training module is used for obtaining an interpolation model by adopting the first sample data and the second equipment for cooperative training, wherein the interpolation model is used for inputting data belonging to the data characteristics corresponding to the first equipment and outputting prediction data belonging to the data characteristics corresponding to the second equipment.
9. A longitudinal federated learning system optimization apparatus, comprising: a memory, a processor, and a longitudinal federated learning system optimization program stored on the memory and executable on the processor, the longitudinal federated learning system optimization program when executed by the processor implementing the steps of the longitudinal federated learning system optimization method of any one of claims 1 to 7.
10. A computer readable storage medium having stored thereon a longitudinal federated learning system optimization program that, when executed by a processor, performs the steps of the longitudinal federated learning system optimization method of any of claims 1-7.
CN201910918262.9A 2019-09-26 2019-09-26 Longitudinal federal learning system optimization method, device, equipment and readable storage medium Active CN110633805B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201910918262.9A CN110633805B (en) 2019-09-26 2019-09-26 Longitudinal federal learning system optimization method, device, equipment and readable storage medium
PCT/CN2020/107332 WO2021004551A1 (en) 2019-09-26 2020-08-06 Method, apparatus, and device for optimization of vertically federated learning system, and a readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910918262.9A CN110633805B (en) 2019-09-26 2019-09-26 Longitudinal federal learning system optimization method, device, equipment and readable storage medium

Publications (2)

Publication Number Publication Date
CN110633805A true CN110633805A (en) 2019-12-31
CN110633805B CN110633805B (en) 2024-04-26

Family

ID=68973220

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910918262.9A Active CN110633805B (en) 2019-09-26 2019-09-26 Longitudinal federal learning system optimization method, device, equipment and readable storage medium

Country Status (2)

Country Link
CN (1) CN110633805B (en)
WO (1) WO2021004551A1 (en)

Cited By (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111241580A (en) * 2020-01-09 2020-06-05 广州大学 Trusted execution environment-based federated learning method
CN111260061A (en) * 2020-03-09 2020-06-09 厦门大学 Differential noise adding method and system in federated learning gradient exchange
CN111310204A (en) * 2020-02-10 2020-06-19 北京百度网讯科技有限公司 Data processing method and device
CN111339212A (en) * 2020-02-13 2020-06-26 深圳前海微众银行股份有限公司 Sample clustering method, device, equipment and readable storage medium
CN111401552A (en) * 2020-03-11 2020-07-10 浙江大学 Federal learning method and system based on batch size adjustment and gradient compression rate adjustment
CN111460511A (en) * 2020-04-17 2020-07-28 支付宝(杭州)信息技术有限公司 Federal learning and virtual object distribution method and device based on privacy protection
CN111625820A (en) * 2020-05-29 2020-09-04 华东师范大学 Federal defense method based on AIoT-oriented security
CN111737921A (en) * 2020-06-24 2020-10-02 深圳前海微众银行股份有限公司 Data processing method, device and medium based on recurrent neural network
CN111858955A (en) * 2020-07-01 2020-10-30 石家庄铁路职业技术学院 Knowledge graph representation learning enhancement method and device based on encrypted federated learning
CN111859360A (en) * 2020-07-10 2020-10-30 深圳市洞见智慧科技有限公司 Safe multi-device joint data computing system, method and device
CN111931216A (en) * 2020-09-16 2020-11-13 支付宝(杭州)信息技术有限公司 Method and system for obtaining joint training model based on privacy protection
CN112149171A (en) * 2020-10-27 2020-12-29 腾讯科技(深圳)有限公司 Method, device, equipment and storage medium for training federal neural network model
WO2021004551A1 (en) * 2019-09-26 2021-01-14 深圳前海微众银行股份有限公司 Method, apparatus, and device for optimization of vertically federated learning system, and a readable storage medium
CN112231746A (en) * 2020-09-10 2021-01-15 杭州锘崴信息科技有限公司 Joint data analysis method, device and system and computer readable storage medium
CN112418446A (en) * 2020-11-18 2021-02-26 脸萌有限公司 Model processing method, system, device, medium and electronic equipment
CN112529102A (en) * 2020-12-24 2021-03-19 深圳前海微众银行股份有限公司 Feature expansion method, device, medium, and computer program product
CN112700010A (en) * 2020-12-30 2021-04-23 深圳前海微众银行股份有限公司 Feature completion method, device, equipment and storage medium based on federal learning
CN112818374A (en) * 2021-03-02 2021-05-18 深圳前海微众银行股份有限公司 Joint training method, device, storage medium and program product of model
CN112836767A (en) * 2021-03-03 2021-05-25 深圳前海微众银行股份有限公司 Federal modeling method, apparatus, device, storage medium, and program product
CN113392164A (en) * 2020-03-13 2021-09-14 京东城市(北京)数字科技有限公司 Method, main server, service platform and system for constructing longitudinal federated tree
CN113554476A (en) * 2020-04-23 2021-10-26 京东数字科技控股有限公司 Training method and system of credit prediction model, electronic device and storage medium
WO2021219054A1 (en) * 2020-04-29 2021-11-04 深圳前海微众银行股份有限公司 Transverse federated learning system optimization method, apparatus and device, and readable storage medium
WO2022016964A1 (en) * 2020-07-23 2022-01-27 深圳前海微众银行股份有限公司 Vertical federated modeling optimization method and device, and readable storage medium
CN114186694A (en) * 2021-11-16 2022-03-15 浙江大学 Efficient, safe and low-communication longitudinal federal learning method
WO2022068575A1 (en) * 2020-09-30 2022-04-07 腾讯科技(深圳)有限公司 Calculation method for vertical federated learning, apparatus, device, and medium
CN114302412A (en) * 2021-12-27 2022-04-08 中国电信股份有限公司 Base station construction intelligent planning method and device, storage medium and terminal equipment
CN114492846A (en) * 2022-04-06 2022-05-13 天聚地合(苏州)科技股份有限公司 Cross-domain federated learning method and system based on trusted execution environment
CN114548255A (en) * 2022-02-17 2022-05-27 支付宝(杭州)信息技术有限公司 Model training method, device and equipment
CN115659212A (en) * 2022-09-27 2023-01-31 南京邮电大学 Federal learning efficiency evaluation method based on TDD communication under cross-domain heterogeneous scene
WO2023024350A1 (en) * 2021-08-25 2023-03-02 深圳前海微众银行股份有限公司 Longitudinal federal prediction optimization method and device, medium, and computer program product
CN116032431A (en) * 2023-02-06 2023-04-28 哈尔滨工业大学(深圳) Communication compression method for distributed learning and related equipment
WO2023134077A1 (en) * 2022-01-17 2023-07-20 平安科技(深圳)有限公司 Homomorphic encryption method and system based on federated factorization machine, device and storage medium
CN117689009A (en) * 2024-01-30 2024-03-12 东北大学 Regularized personalized federal training-oriented communication optimization method

Families Citing this family (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112686473B (en) * 2021-01-22 2023-12-08 国网江苏省电力有限公司营销服务中心 Power consumption prediction method and prediction system based on classification algorithm
CN112885337A (en) * 2021-01-29 2021-06-01 深圳前海微众银行股份有限公司 Data processing method, device, equipment and storage medium
CN112818394A (en) * 2021-01-29 2021-05-18 西安交通大学 Self-adaptive asynchronous federal learning method with local privacy protection
CN112989944A (en) * 2021-02-08 2021-06-18 西安翔迅科技有限责任公司 Intelligent video safety supervision method based on federal learning
CN112948883B (en) * 2021-03-25 2023-10-31 支付宝(杭州)信息技术有限公司 Method, device and system for multiparty joint modeling of privacy data protection
CN113191072A (en) * 2021-03-30 2021-07-30 中国建设银行股份有限公司 Suspicious transaction monitoring method and device based on longitudinal federal logistic regression
CN112949760B (en) * 2021-03-30 2024-05-10 平安科技(深圳)有限公司 Model precision control method, device and storage medium based on federal learning
CN113807535B (en) * 2021-04-01 2023-11-03 京东科技控股股份有限公司 Training method and device of federal learning model, electronic equipment and storage medium
CN113807538B (en) * 2021-04-09 2024-02-06 京东科技控股股份有限公司 Federal learning method, federal learning device, electronic equipment and storage medium
CN113269232B (en) * 2021-04-25 2023-12-08 北京沃东天骏信息技术有限公司 Model training method, vectorization recall method, related equipment and storage medium
CN113240461B (en) * 2021-05-07 2022-08-16 广州银行股份有限公司 Method, system and medium for identifying potential customers based on longitudinal federal learning
CN113283521B (en) * 2021-06-03 2023-09-01 光大科技有限公司 Condition generation countermeasure network generation method and device
CN113361771B (en) * 2021-06-04 2023-04-18 合肥工业大学 Method and device for determining the pressure of a storage tank
CN113297573B (en) * 2021-06-11 2022-06-10 浙江工业大学 Vertical federal learning defense method and device based on GAN simulation data generation
CN113642034A (en) * 2021-06-25 2021-11-12 合肥工业大学 Medical big data safety sharing method and system based on horizontal and vertical federal learning
CN113642664B (en) * 2021-08-24 2024-02-20 安徽大学 Privacy protection image classification method based on federal learning
CN113688771B (en) * 2021-09-02 2024-04-05 重庆大学 LNG storage tank acceleration response data supplementing method and device
CN114186263B (en) * 2021-12-17 2024-05-03 大连理工大学 Data regression method based on longitudinal federal learning and electronic device
CN114429223B (en) * 2022-01-26 2023-11-07 上海富数科技有限公司 Heterogeneous model building method and device
CN114818476B (en) * 2022-04-01 2023-08-22 西南交通大学 Federal learning system and method applied to life prediction of rotating mechanical equipment
CN114996317B (en) * 2022-07-05 2024-02-23 中国电信股份有限公司 Asynchronous optimization method and device based on longitudinal federal learning and storage medium
CN115292738B (en) * 2022-10-08 2023-01-17 豪符密码检测技术(成都)有限责任公司 Method for detecting security and correctness of federated learning model and data
CN115545216B (en) * 2022-10-19 2023-06-30 上海零数众合信息科技有限公司 Service index prediction method, device, equipment and storage medium
CN115563564B (en) * 2022-12-02 2023-03-17 腾讯科技(深圳)有限公司 Processing method and device of decision tree model, computer equipment and storage medium
CN116319714B (en) * 2023-05-24 2023-07-21 北京邮电大学 Federal learning method based on modal conversion and related equipment
CN116633704B (en) * 2023-07-25 2023-10-31 北京数牍科技有限公司 Graph calculation method and device
CN117094421B (en) * 2023-10-16 2023-12-15 亚信科技(中国)有限公司 Asymmetric longitudinal federal learning method, device, electronic equipment and storage medium
CN117853212B (en) * 2024-03-06 2024-06-18 之江实验室 Longitudinal federal financial wind control method based on knowledge migration and self-supervision learning

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100076785A1 (en) * 2008-09-25 2010-03-25 Air Products And Chemicals, Inc. Predicting rare events using principal component analysis and partial least squares
US20180322365A1 (en) * 2017-05-05 2018-11-08 Intel Corporation On the fly deep learning in machine learning for autonomous machines
CN109347620A (en) * 2018-08-10 2019-02-15 深圳前海微众银行股份有限公司 Sample alignment schemes, system and computer readable storage medium

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109033854B (en) * 2018-07-17 2020-06-09 阿里巴巴集团控股有限公司 Model-based prediction method and device
CN109002861B (en) * 2018-08-10 2021-11-09 深圳前海微众银行股份有限公司 Federal modeling method, device and storage medium
CN110110229B (en) * 2019-04-25 2021-06-04 深圳前海微众银行股份有限公司 Information recommendation method and device
CN110245510B (en) * 2019-06-19 2021-12-07 北京百度网讯科技有限公司 Method and apparatus for predicting information
CN110633805B (en) * 2019-09-26 2024-04-26 深圳前海微众银行股份有限公司 Longitudinal federal learning system optimization method, device, equipment and readable storage medium

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100076785A1 (en) * 2008-09-25 2010-03-25 Air Products And Chemicals, Inc. Predicting rare events using principal component analysis and partial least squares
US20180322365A1 (en) * 2017-05-05 2018-11-08 Intel Corporation On the fly deep learning in machine learning for autonomous machines
CN109347620A (en) * 2018-08-10 2019-02-15 深圳前海微众银行股份有限公司 Sample alignment schemes, system and computer readable storage medium

Cited By (51)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021004551A1 (en) * 2019-09-26 2021-01-14 深圳前海微众银行股份有限公司 Method, apparatus, and device for optimization of vertically federated learning system, and a readable storage medium
CN111241580A (en) * 2020-01-09 2020-06-05 广州大学 Trusted execution environment-based federated learning method
CN111241580B (en) * 2020-01-09 2022-08-09 广州大学 Trusted execution environment-based federated learning method
CN111310204A (en) * 2020-02-10 2020-06-19 北京百度网讯科技有限公司 Data processing method and device
CN111310204B (en) * 2020-02-10 2022-06-14 北京百度网讯科技有限公司 Data processing method and device
CN111339212A (en) * 2020-02-13 2020-06-26 深圳前海微众银行股份有限公司 Sample clustering method, device, equipment and readable storage medium
CN111260061A (en) * 2020-03-09 2020-06-09 厦门大学 Differential noise adding method and system in federated learning gradient exchange
CN111260061B (en) * 2020-03-09 2022-07-19 厦门大学 Differential noise adding method and system in federated learning gradient exchange
CN111401552B (en) * 2020-03-11 2023-04-07 浙江大学 Federal learning method and system based on batch size adjustment and gradient compression rate adjustment
CN111401552A (en) * 2020-03-11 2020-07-10 浙江大学 Federal learning method and system based on batch size adjustment and gradient compression rate adjustment
CN113392164B (en) * 2020-03-13 2024-01-12 京东城市(北京)数字科技有限公司 Method for constructing longitudinal federal tree, main server, service platform and system
CN113392164A (en) * 2020-03-13 2021-09-14 京东城市(北京)数字科技有限公司 Method, main server, service platform and system for constructing longitudinal federated tree
CN111460511A (en) * 2020-04-17 2020-07-28 支付宝(杭州)信息技术有限公司 Federal learning and virtual object distribution method and device based on privacy protection
CN111460511B (en) * 2020-04-17 2023-05-02 支付宝(杭州)信息技术有限公司 Federal learning and virtual object distribution method and device based on privacy protection
CN113554476A (en) * 2020-04-23 2021-10-26 京东数字科技控股有限公司 Training method and system of credit prediction model, electronic device and storage medium
CN113554476B (en) * 2020-04-23 2024-04-19 京东科技控股股份有限公司 Training method and system of credit prediction model, electronic equipment and storage medium
WO2021219054A1 (en) * 2020-04-29 2021-11-04 深圳前海微众银行股份有限公司 Transverse federated learning system optimization method, apparatus and device, and readable storage medium
CN111625820A (en) * 2020-05-29 2020-09-04 华东师范大学 Federal defense method based on AIoT-oriented security
CN111737921A (en) * 2020-06-24 2020-10-02 深圳前海微众银行股份有限公司 Data processing method, device and medium based on recurrent neural network
CN111737921B (en) * 2020-06-24 2024-04-26 深圳前海微众银行股份有限公司 Data processing method, equipment and medium based on cyclic neural network
CN111858955B (en) * 2020-07-01 2023-08-18 石家庄铁路职业技术学院 Knowledge graph representation learning enhancement method and device based on encryption federal learning
CN111858955A (en) * 2020-07-01 2020-10-30 石家庄铁路职业技术学院 Knowledge graph representation learning enhancement method and device based on encrypted federated learning
CN111859360A (en) * 2020-07-10 2020-10-30 深圳市洞见智慧科技有限公司 Safe multi-device joint data computing system, method and device
WO2022016964A1 (en) * 2020-07-23 2022-01-27 深圳前海微众银行股份有限公司 Vertical federated modeling optimization method and device, and readable storage medium
CN112231746B (en) * 2020-09-10 2024-02-02 杭州锘崴信息科技有限公司 Joint data analysis method, device, system and computer readable storage medium
CN112231746A (en) * 2020-09-10 2021-01-15 杭州锘崴信息科技有限公司 Joint data analysis method, device and system and computer readable storage medium
US11270028B1 (en) 2020-09-16 2022-03-08 Alipay (Hangzhou) Information Technology Co., Ltd. Obtaining jointly trained model based on privacy protection
CN111931216A (en) * 2020-09-16 2020-11-13 支付宝(杭州)信息技术有限公司 Method and system for obtaining joint training model based on privacy protection
WO2022068575A1 (en) * 2020-09-30 2022-04-07 腾讯科技(深圳)有限公司 Calculation method for vertical federated learning, apparatus, device, and medium
CN112149171B (en) * 2020-10-27 2021-07-09 腾讯科技(深圳)有限公司 Method, device, equipment and storage medium for training federal neural network model
CN112149171A (en) * 2020-10-27 2020-12-29 腾讯科技(深圳)有限公司 Method, device, equipment and storage medium for training federal neural network model
CN112418446A (en) * 2020-11-18 2021-02-26 脸萌有限公司 Model processing method, system, device, medium and electronic equipment
CN112418446B (en) * 2020-11-18 2024-04-09 脸萌有限公司 Model processing method, system, device, medium and electronic equipment
CN112529102A (en) * 2020-12-24 2021-03-19 深圳前海微众银行股份有限公司 Feature expansion method, device, medium, and computer program product
CN112529102B (en) * 2020-12-24 2024-03-12 深圳前海微众银行股份有限公司 Feature expansion method, device, medium and computer program product
CN112700010A (en) * 2020-12-30 2021-04-23 深圳前海微众银行股份有限公司 Feature completion method, device, equipment and storage medium based on federal learning
CN112818374A (en) * 2021-03-02 2021-05-18 深圳前海微众银行股份有限公司 Joint training method, device, storage medium and program product of model
CN112836767A (en) * 2021-03-03 2021-05-25 深圳前海微众银行股份有限公司 Federal modeling method, apparatus, device, storage medium, and program product
WO2023024350A1 (en) * 2021-08-25 2023-03-02 深圳前海微众银行股份有限公司 Longitudinal federal prediction optimization method and device, medium, and computer program product
CN114186694A (en) * 2021-11-16 2022-03-15 浙江大学 Efficient, safe and low-communication longitudinal federal learning method
CN114186694B (en) * 2021-11-16 2024-06-11 浙江大学 Efficient, safe and low-communication longitudinal federal learning method
CN114302412A (en) * 2021-12-27 2022-04-08 中国电信股份有限公司 Base station construction intelligent planning method and device, storage medium and terminal equipment
WO2023134077A1 (en) * 2022-01-17 2023-07-20 平安科技(深圳)有限公司 Homomorphic encryption method and system based on federated factorization machine, device and storage medium
CN114548255A (en) * 2022-02-17 2022-05-27 支付宝(杭州)信息技术有限公司 Model training method, device and equipment
CN114492846B (en) * 2022-04-06 2022-08-26 天聚地合(苏州)科技股份有限公司 Cross-domain federated learning method and system based on trusted execution environment
CN114492846A (en) * 2022-04-06 2022-05-13 天聚地合(苏州)科技股份有限公司 Cross-domain federated learning method and system based on trusted execution environment
CN115659212B (en) * 2022-09-27 2024-04-09 南京邮电大学 Federal learning efficiency evaluation method based on TDD communication under cross-domain heterogeneous scene
CN115659212A (en) * 2022-09-27 2023-01-31 南京邮电大学 Federal learning efficiency evaluation method based on TDD communication under cross-domain heterogeneous scene
CN116032431B (en) * 2023-02-06 2023-09-12 哈尔滨工业大学(深圳) Communication compression method for distributed learning and related equipment
CN116032431A (en) * 2023-02-06 2023-04-28 哈尔滨工业大学(深圳) Communication compression method for distributed learning and related equipment
CN117689009A (en) * 2024-01-30 2024-03-12 东北大学 Regularized personalized federal training-oriented communication optimization method

Also Published As

Publication number Publication date
CN110633805B (en) 2024-04-26
WO2021004551A1 (en) 2021-01-14

Similar Documents

Publication Publication Date Title
CN110633805B (en) Longitudinal federal learning system optimization method, device, equipment and readable storage medium
CN110633806B (en) Longitudinal federal learning system optimization method, device, equipment and readable storage medium
CN109492420B (en) Model parameter training method, terminal, system and medium based on federal learning
US11934540B2 (en) System and method for multiparty secure computing platform
US11743038B2 (en) Methods and systems of providing verification of information using a centralized or distributed ledger
US11121857B2 (en) Systems, devices, and methods for in-field authenticating of autonomous robots
CN110535833B (en) Data sharing control method based on block chain
US20200403992A1 (en) Universal Digital Identity Authentication Service
US20210406386A1 (en) System and method for multiparty secure computing platform
US20170046693A1 (en) Systems and methods for detecting and resolving data inconsistencies among networked devices using hybrid private-public blockchain ledgers
US20190222418A1 (en) Systems and Methods for Key Exchange in Blockchain
CN109446259B (en) Data processing method and device, processor and storage medium
US20190098013A1 (en) System and Methods for Location Verification with Blockchain Controls
US20190288833A1 (en) System and Method for Securing Private Keys Behind a Biometric Authentication Gateway
CN116992458A (en) Programmable data processing method and system based on trusted execution environment
US11943210B2 (en) System and method for distributed, keyless electronic transactions with authentication
Althati et al. Scalable Machine Learning Solutions for Heterogeneous Data in Distributed Data Platform
CN116502732A (en) Federal learning method and system based on trusted execution environment
CN115270159A (en) Intelligent contract calling method, device and equipment for block chain and storage medium
CN210745178U (en) Identity authentication system
TWM585941U (en) Account data processing system
CN111125734A (en) Data processing method and system
TW202101267A (en) Account data processing method and account data processing system ensuring that there is encryption protection when account data is returned to an electronic payment dealer
CN115510504B (en) Data sharing method, system, equipment and medium based on ring signature and promise
Zeydan et al. Blockchain-based self-sovereign identity solution for aerial base station integrated networks

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant