CN112634027A - Self-adaptive federal parameter aggregation method for credit assessment of small and micro enterprises - Google Patents
Self-adaptive federal parameter aggregation method for credit assessment of small and micro enterprises Download PDFInfo
- Publication number
- CN112634027A CN112634027A CN202011600934.0A CN202011600934A CN112634027A CN 112634027 A CN112634027 A CN 112634027A CN 202011600934 A CN202011600934 A CN 202011600934A CN 112634027 A CN112634027 A CN 112634027A
- Authority
- CN
- China
- Prior art keywords
- data
- parameter
- similarity
- vector
- matrix
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 46
- 230000002776 aggregation Effects 0.000 title claims abstract description 29
- 238000004220 aggregation Methods 0.000 title claims abstract description 29
- 239000011159 matrix material Substances 0.000 claims abstract description 56
- 238000012549 training Methods 0.000 claims abstract description 49
- 238000001514 detection method Methods 0.000 claims abstract description 25
- 239000013598 vector Substances 0.000 claims description 63
- 238000005070 sampling Methods 0.000 claims description 9
- 230000003044 adaptive effect Effects 0.000 claims description 8
- 230000007423 decrease Effects 0.000 claims description 2
- 238000003062 neural network model Methods 0.000 claims description 2
- 230000004931 aggregating effect Effects 0.000 claims 7
- 230000008569 process Effects 0.000 description 4
- 238000010801 machine learning Methods 0.000 description 3
- 239000000463 material Substances 0.000 description 3
- 230000007246 mechanism Effects 0.000 description 3
- 238000013526 transfer learning Methods 0.000 description 3
- 238000004891 communication Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 101100268668 Caenorhabditis elegans acc-2 gene Proteins 0.000 description 1
- 101100268670 Caenorhabditis elegans acc-3 gene Proteins 0.000 description 1
- 101100373202 Rattus norvegicus Cx3cl1 gene Proteins 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000033228 biological regulation Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000013508 migration Methods 0.000 description 1
- 230000005012 migration Effects 0.000 description 1
- 230000008092 positive effect Effects 0.000 description 1
- 230000029305 taxis Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q40/00—Finance; Insurance; Tax strategies; Processing of corporate or income taxes
- G06Q40/03—Credit; Loans; Processing thereof
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/60—Protecting data
- G06F21/62—Protecting access to data via a platform, e.g. using keys or access control rules
- G06F21/6218—Protecting access to data via a platform, e.g. using keys or access control rules to a system of files or objects, e.g. local or distributed file system or database
- G06F21/6245—Protecting personal data, e.g. for financial or medical purposes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Business, Economics & Management (AREA)
- Software Systems (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Accounting & Taxation (AREA)
- Finance (AREA)
- Medical Informatics (AREA)
- General Health & Medical Sciences (AREA)
- Bioethics (AREA)
- Databases & Information Systems (AREA)
- Marketing (AREA)
- Economics (AREA)
- General Business, Economics & Management (AREA)
- Strategic Management (AREA)
- Computer Hardware Design (AREA)
- Computer Security & Cryptography (AREA)
- Development Economics (AREA)
- Technology Law (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The invention discloses a self-adaptive federal parameter aggregation method for credit assessment of a small micro-enterprise, which specifically comprises the following steps: s1, constructing a network by the participants; s2, training the participants locally to obtain accuracy; s3, carrying out similarity detection on the participants by using the intelligent contract; s4, obtaining a training matrix and selecting participants to perform parameter aggregation; s5, returning the parameters to the participants, and retraining to obtain the accuracy; s6, judging the accuracy of the two times, and if the accuracy is improved, ending the training; otherwise, updating the weight and the aggregated parameters in the distance matrix; s7, carrying out parameter aggregation again according to the training matrix; and S8, until all participants are improved in result. The method can automatically select the participants for parameter aggregation, does not need to perform parameter aggregation on all the participants, adaptively updates the weight in the distance matrix according to the accuracy before and after training, does not need third party participation, is decentralized, and can ensure the privacy of data information.
Description
Technical Field
The invention relates to the field of credit assessment of small and micro enterprises, in particular to a self-adaptive federal parameter aggregation method for credit assessment of small and micro enterprises.
Background
A common method currently used to address data islanding and data privacy is federal learning. Federal learning was originally proposed by ***, and their main idea was to build machine learning models based on data sets distributed across multiple devices while preventing data leakage. Recent improvements have focused on overcoming statistical challenges and improving security for federal learning. The Oerson bank is a mechanism which provides a 'federal study' which is a general solution for solving the problems of data island and data privacy protection in artificial intelligence landing and is provided by the first family in China.
Federal learning is a cryptographic distributed machine learning technique that allows users to train machine learning models using multiple data sets distributed over different locations, while preventing data leakage and adhering to strict data privacy regulations. Federal learning is a learning process in which data owners jointly train a model while returning trained parameters to the participants. In this process, any owner of the data will not expose its data to others.
For different data sets, the federal learning is divided into horizontal federal learning, vertical federal learning and federal transfer learning. In the case of more user feature overlap and less user overlap of two data sets, we divide the data sets according to the horizontal direction (i.e. user dimension), and extract the part of data with the same user feature but not identical user for training. This method is called horizontal federal learning. Longitudinal federated learning under the condition that users of two data sets overlap more and user features overlap less, the data sets are divided according to the longitudinal direction (namely feature dimension), and the data of which the users are the same and the user features are not completely the same is taken out for training. This method is called longitudinal federal learning. Federal transfer learning under the condition that users of two data sets are less in feature overlapping with the users, data are not segmented, and the condition that data or labels are insufficient is overcome by transfer learning. This method is called federal migration learning.
At present, federal learning adopts a centralized aggregation method, all participant data need to be aggregated into parameters, each party of data participating in aggregation cannot be ensured to have positive effect, and the problem of large communication burden exists in a centralized parameter aggregation process. In fact, a federated network may be composed of a large number of devices, such as millions of smartphones, and the communication speed in the network may be many orders of magnitude slower than local computing.
Disclosure of Invention
The invention aims to provide a self-adaptive federal parameter aggregation method for small-scale enterprise credit assessment, which aims to solve the problems in the prior art and achieve the best accuracy of credit assessment for small-scale enterprises.
In order to achieve the purpose, the invention provides the following scheme:
the invention provides a self-adaptive federal parameter aggregation method for credit assessment of small micro-enterprises, which comprises the following specific steps:
s1, acquiring the geographical position information of the participants, constructing a network by taking the geographical position information as a data set, and calculating a distance matrix of the nodes of the participants;
s2, constructing a training model, and locally training the position information of the participants to obtain the accuracy;
s3, carrying out similarity detection (divided into data and parameter similarity detection) on the participants by using the intelligent contract;
s4, obtaining a training matrix selection model and carrying out parameter aggregation;
s5, returning the parameters to the model, and retraining to obtain the accuracy;
s6, judging the accuracy of the two times, simultaneously updating the weight in the aggregation parameter,
s7, if the accuracy is improved, the training is finished; otherwise, updating the weight and the aggregated parameters in the distance matrix, and performing parameter aggregation again according to the training matrix;
and S8, until all participants are improved in result.
The detailed process is as follows:
and S1, building a network according to the geographical position of the participant to calculate a distance matrix A for the participant nodes, wherein the distance matrix is a matrix representing the distance relationship between the nodes and is similar to the definition of the adjacent matrix of the undirected graph. Assume that there are n participants, i.e., n nodes.
The distance matrix is defined as:
wherein wijFor weight, the weight is great when the distance between the nodes i and j is close, and the weight is small when the distance is far, A[i,j]The distance value between the nodes i and j corresponds to a value in the distance matrix, and k is the number of the nodes between the nodes i and j;
s2, constructing a neural network model for each participantThe person performs model training locally to obtain the accuracy A of individual trainingcci;
And S3, carrying out similarity detection on the participants by using the intelligent contract: the method comprises data similarity detection and parameter similarity detection.
1. And (3) detecting the data similarity:
vectorizing data, specifically including:
segmenting data, converting characters into one dimension in vector by word embedding, embedding words by combining numbers with words in front and back, and embedding words by Qi=τiQi-1+τi+1Qi+1Wherein Q isiEmbedding vectors, Q, for digital wordsi-1One-dimensional, Q, of the transformed vector for the word preceding the numberi+1One-dimensional, tau, of the transformed vector of the word after the numberiAnd τi+1Are all weights, let τ beiIs set to 0.6, taui+1Is set to 0.4, a vector E of a segment is obtained1Each segment is then vectorized. Finally according toObtaining a data vector, and then carrying out data similarity detection;
carrying out data similarity detection, specifically comprising:
any two participants a and b are taken as vectors x and y after data vectorization, and x is assumed to be [ x ═ x [1,x2…xn]T,y=[y1,y2…yn]T。
1) a, b inform each other of the modular length of the respective data vector, i.e. x1And y1And simultaneously a informs b of any hyperplane passing x: z is u1z1+u2z2+…+unznC, randomly selecting one of the hyperplanes which pass x, informing b, wherein c is the hyperplane definition, and the value is a constant;
2) b knowing a hyperplane in a, calculating the projection vector y of y on z||And informs a. Due to the hyperplane methodThe vector is n ═ n1,n2…nN]TThen the projection vector of y on n is Thus is provided with
3) a learning y||Calculating the distance of similarityAnd inform b of epsiloniFor weighting, the initial value is set to 1, since y ═ y||+y⊥And x y⊥0, so x y||Then, reselecting a participant data vector, and calculating the similarity distance to obtain a data similarity matrix:
and (3) mechanism analysis:
1) a only knows the length y of y1And y projection vector y on z||Y cannot be completely recovered;
2) b knowing only the length x of x1And a hyperplane z passing x, x cannot be completely recovered;
3) the privacy of the vector information can be ensured without the participation of a third party and decentralization.
2. Parameter similarity detection
Vectorizing the model parameters: and sequentially sampling parameters of each layer in the model trained by the participant according to a sampling rejection method, wherein the sampled parameters are used as one dimension of a parameter vector to obtain the parameter vector of the model, and the dimension with less layers is automatically supplemented by 0 to be consistent with the dimension with the length of the vector.
Detecting the parameter similarity: and (4) randomly taking the parameter vectors of the c model and the d model, and calculating the similarity distance, wherein the formula is shown as follows. And reselecting the two parameter vectors, and carrying out parameter similarity detection until all the models are detected, and finally obtaining a parameter similarity matrix.
s3, combining the data similarity matrix and the parameter similarity matrix to obtain a training matrix: training matrixWherein T is a training matrix, P is a data similarity matrix, S is a parameter similarity matrix, x is a value between 0 and 4, the distance of the value is set to be 0.001, the distance of the value is set according to the size of a participant, the weight of the data similarity is large when the participant starts, and then the data similarity gradually decreases;
s4, selecting a model with a high value in the training matrix to carry out parameter aggregation according to the training requirement, and carrying out parameter calculation by adopting the following formula;
wherein sigmaiAs the weight, the initial value is set to 1;
s5, returning the aggregated parameters to the model, retraining and recording the accuracy Accj;
S6, judgment Accj>AcciUpdating σ in aggregated parameters simultaneouslyi:
Wherein sigmaiAn initial value is set to 1 for the weight, parameters are updated according to the result of training,the weights after the update. Assuming that the training of the two models of c and d is finished, then obtaining the next training sigma of the model c according to the formulaiTaking the weight as the weight for parameter aggregation of c and other models;
s7, if Accj>AcciThen the selection model training is finished. Otherwise, updating w of the training model in the distance matrixij:Wherein eta is the update step length,are the weights in the updated distance matrix. Updating the corresponding w in the distance matrixijCarrying out similarity detection again to obtain an updated training matrix, and reselecting a model with a high value in the training matrix for training until the accuracy is improved;
s8, reselecting the participants for training (excluding the model which is trained before);
s8, all model results are improved.
The invention discloses the following technical effects:
the method can automatically select the participants to perform parameter aggregation without parameter aggregation of all the participants, the parameters and the data similarity are combined to select the participants to perform parameter aggregation, meanwhile, the participants are reselected to perform parameter aggregation according to the weight in the distance matrix updated in a self-adaptive manner before and after training, and the method does not need participation of a third party, is decentralized, and can ensure the privacy of vector information and data information.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings needed in the embodiments will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings without creative efforts.
Fig. 1 is a flowchart of an adaptive federal parameter aggregation method for small micro-enterprise credit assessment.
Detailed Description
Reference will now be made in detail to various exemplary embodiments of the invention, the detailed description should not be construed as limiting the invention but as a more detailed description of certain aspects, features and embodiments of the invention.
It is to be understood that the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. Further, for numerical ranges in this disclosure, it is understood that each intervening value, between the upper and lower limit of that range, is also specifically disclosed. Every smaller range between any stated value or intervening value in a stated range and any other stated or intervening value in a stated range is encompassed within the invention. The upper and lower limits of these smaller ranges may independently be included or excluded in the range.
Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. Although only preferred methods and materials are described herein, any methods and materials similar or equivalent to those described herein can be used in the practice or testing of the present invention. All documents mentioned in this specification are incorporated by reference herein for the purpose of disclosing and describing the methods and/or materials associated with the documents. In case of conflict with any incorporated document, the present specification will control.
It will be apparent to those skilled in the art that various modifications and variations can be made in the specific embodiments of the present disclosure without departing from the scope or spirit of the disclosure. Other embodiments will be apparent to those skilled in the art from consideration of the specification. The specification and examples are exemplary only.
As used herein, the terms "comprising," "including," "having," "containing," and the like are open-ended terms that mean including, but not limited to.
The "parts" in the present invention are all parts by mass unless otherwise specified.
Assuming that a small micro-enterprise S applies for loan from a business bank, there are now a credit A in the central bank, a business bank B, and a tax bureau C (the example is for brevity, and additional participants, such as agricultural banks, tax bureaus, statistics bureaus, etc.)
1. A network is first constructed based on the geographic location of A, B, C. Assuming the network is A-B-C, according to formula A[xi,yi]=wij(n-k) obtaining the distance of each node, wherein wijFor weight, the weight of the network node near is great, the weight of the node far is small, w is between A and B, and B and CijSet to 2, set to 1 between A and C, then calculate the distance between every two nodes, set to 0 node to node) get the distance matrix
2, A, B and C are trained locally to obtain accuracy acc of each evaluation on company credit1,acc2,acc3Each having recorded training parameters of A: alphaA=0.13,βA=0.53,γA=0.32;B:αB=0.26,βB=0.85,γB=0.64;C:αC=0.42,βC=1.65,γC=1.84
3. The similarity of participant data and parameters is detected using the intelligent contract as follows:
1) and (3) detecting the data similarity: the method is characterized in that the data of 50 ten thousand of deposits in an industrial and commercial bank and 5 ten thousand of taxes per year of C in the enterprise with good credit in the A are segmented and vectorized, and the specific method comprises the following steps:
segmenting data, converting characters into one dimension in vector by word embedding, embedding words by combining numbers with preceding and following words, and embedding words by Qi=τiQi-1+τi+1Qi+1Wherein Q isiEmbedding vectors, Q, for digital wordsi-1One-dimensional, Q, of the transformed vector for the word preceding the numberi+1One-dimensional, tau, of the transformed vector of the word after the numberiAnd τi+1Are all weights, let τ beiIs set to 0.6, taui+1Is set to 0.4, a vector E of a segment is obtained1Then, each segment is vectorized as described above, assuming that three segment vectors of A are obtainedThe vectors are generated from the same dictionary, so the dimensions are the same. Finally, obtaining the data vector of AB data vector is obtained in the same way
And after the data vector is obtained, carrying out data similarity detection on the data:
a, B inform each other of the modular length of the respective data vector, i.e. Same | EB|=2.73,|EC0.74, while A tells B any one of EAThe hyperplane of (a): z is u1z1+u2z2+…+unznWhen it is 0, assume thatWherein x-passing hyperplanes are numerous, and one hyperplane is randomly selectedInforming the user of the information B;
(B) knowing a hyperplane in A, calculate EBProjection vector y on z||And informs a. EBThe projection vector on n is
(iii) A knows y||The similarity distance is calculated because EB ═ y||+y⊥And E isA*y⊥If equal to 0, there is EA*EB=EA*y||。Then calculate P(A,C)=0.62,P(B,C)=0.53
Fourthly, reselecting the data vector of the participant and calculating the similarity distance P(A,C)=0.62,P(B,C)=0.53
And (3) mechanism analysis:
(a) knowing only the length y of y1And y projection vector y on z||Y cannot be completely recovered;
b knowing only the length x of x1And a hyperplane z passing x, x cannot be completely recovered;
and thirdly, the privacy of the vector information can be ensured without the participation of a third party and without centralization.
2) Detecting the parameter similarity: and sequentially sampling trainable parameters of each layer of the model trained by the participants A, B and C according to a sampling rejection method, wherein the trainable parameters are used as one dimension of a parameter vector. And obtaining a parameter vector of the model, and automatically filling the dimension with less layer number to make the dimension consistent with the dimension with the length of the vector. Suppose A trained modelThe number of layers is 2, the first layer has 5 trainable parameters, the second layer has 7 trainable parameters, the model trained by B has 3 layers, the first layer has 6 trainable parameters, the second layer has 6 trainable parameters, and the third layer has 9 trainable parameters. A reject-sampling method is used for sampling each layer of a and B, assuming 3 trainable parameters, since B has one more layer, the parameter vector of a has one dimension of 0. Obtaining A parameter vectorAlso, the same applies toDetecting the similarity of the parameter vectors to obtainThen, d (a, C) is calculated to be 0.23 and d (B, C) is calculated to be 0.56. Obtaining a similarity matrix
4. Assuming A is necessary, then 1.13<1.36, A and C are selected for parameter aggregation
5. Returning the aggregated parameters to A, B and C for retraining, simultaneously recording the accuracy Acc1, Acc2 and Acc3 respectively, and updating the aggregated parameters Wherein sigmaiFor the weights, the initial value is set to 1,the weights after the update. And updating the parameters according to the training result. Assuming that the training of the a model and the b model is finished, then obtaining the next training sigma of the a model according to the formulaiThis is used as a weight for parameter aggregation with other models.
6. If Accj is>Acci, then training of A and C is finished. Otherwise, updating w of the training model in the distance matrixij:Wherein eta is the update step length,are the weights in the updated distance matrix. Updating the corresponding w in the distance matrixijAnd carrying out similarity detection again to obtain an updated training matrix, and reselecting a model with a high value in the training matrix for training until the accuracy is improved.
7. Reselect participants A and B for training (exclude previously trained A and C)
8. All participants trained locally with improved accuracy of the assessment.
The above-described embodiments are merely illustrative of the preferred embodiments of the present invention, and do not limit the scope of the present invention, and various modifications and improvements of the technical solutions of the present invention can be made by those skilled in the art without departing from the spirit of the present invention, and the technical solutions of the present invention are within the scope of the present invention defined by the claims.
Claims (8)
1. A self-adaptive federal parameter aggregation method for small micro-enterprise credit assessment is characterized in that: the method comprises the following steps:
s1, acquiring the geographical position information of the participants, constructing a network by taking the geographical position information as a data set, and calculating a distance matrix of the nodes of the participants;
s2, constructing a neural network model, and carrying out model training on each participant locally to obtain the accuracy of individual training;
and S3, carrying out similarity detection on the participants by using the intelligent contract: the method comprises the steps of dividing data similarity detection and parameter similarity detection, obtaining a data similarity matrix according to the data similarity detection, obtaining a parameter similarity matrix according to the parameter similarity detection, and obtaining a training matrix by combining the data similarity matrix and the parameter similarity matrix;
s4, selecting a model with a high value in the training matrix for parameter aggregation after the training matrix is obtained;
s5, returning the aggregated parameters to the model, retraining and recording the accuracy;
s6, judging the accuracy of the two times, updating the weight in the aggregation parameter, and ending the training if the accuracy is improved; otherwise, updating the weight and the aggregated parameters in the distance matrix, and carrying out similarity detection again;
s7, carrying out parameter aggregation again according to the training matrix;
and S8, until all participants are improved in result.
2. The method for aggregating adaptive federal parameters for small micro-enterprise credit assessment according to claim 1, wherein the method comprises the following steps: the data similarity detection in S3 includes data vectorization and data similarity detection, where the data vectorization is: segmenting data, converting characters into one dimension in vector by word embedding, embedding words by combining numbers with words in front and back, and embedding words by QiObtaining a vector of a segment by a formula, then vectorizing each segment, and finally obtaining a vector of a segment according to the formulaCarrying out data similarity detection after obtaining data vectors, wherein E is the total data vector, EiIs a vector for each segment.
3. The method for aggregating adaptive federal parameters for small micro-enterprise credit assessment according to claim 2, wherein the method comprises the following steps: the data similarity detection comprises: vectors x and y after data vectorization of any two participants a and b are taken;
1) a and b inform each other of the modular length of the respective data vector, and a informs b of any hyperplane z passing x;
2) b, knowing a hyperplane in a, calculating a projection vector of y on z and informing a, knowing a normal vector n of the hyperplane, and obtaining a projection vector of y on n;
3) and a, acquiring a projection vector of y on z, calculating the similarity distance and informing b, then reselecting a participant data vector, and calculating the similarity distance to obtain a data similarity matrix.
4. The method for aggregating adaptive federal parameters for small micro-enterprise credit assessment according to claim 2, wherein the method comprises the following steps: said QiIs given by the formula Qi=τiQi-1+τi+1Qi+1Wherein Q isiEmbedding vectors, Q, for digital wordsi-1One-dimensional, Q, of the transformed vector for the word preceding the numberi+1One-dimensional, tau, of the transformed vector of the word after the numberiAnd τi+1Are all weights, let τ beiIs set to 0.6, taui+1Is set to 0.4.
5. The method for aggregating adaptive federal parameters for small micro-enterprise credit assessment according to claim 1, wherein the method comprises the following steps: the parameter similarity detection in S3 includes model parameter vectorization and parameter similarity detection, where the model parameter vectorization is: and sequentially sampling parameters of each layer in the model trained by the participant, wherein the sampled parameters are used as one dimension of a parameter vector to obtain the parameter vector of the model, and the dimension with less layers is automatically supplemented by 0 to be consistent with the dimension with the length of the vector.
6. The method for aggregating the adaptive federal parameters for small micro-enterprise credit assessment according to claim 5, wherein: the sampling is performed according to a rejection sampling method.
7. The method for aggregating adaptive federal parameters for small micro-enterprise credit assessment according to claim 1, wherein the method comprises the following steps: the parameter similarity in S3 is detected as: and (4) randomly taking the parameter vectors of the c model and the d model, calculating the similarity distance, reselecting the two parameter vectors, detecting the parameter similarity until all the models are detected, and finally obtaining a parameter similarity matrix.
8. The method for aggregating adaptive federal parameters for small micro-enterprise credit assessment according to claim 1, wherein the method comprises the following steps: the training matrix in S3 is:
wherein T is a training matrix, P is a data similarity matrix, S is a parameter similarity matrix, x is taken from 0 to 4, the distance between the values is set to be 0.001, the distance between the values is set according to the size of participants, the weight of the data similarity is large when the data similarity is started, and then the data similarity gradually decreases.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011600934.0A CN112634027A (en) | 2020-12-30 | 2020-12-30 | Self-adaptive federal parameter aggregation method for credit assessment of small and micro enterprises |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011600934.0A CN112634027A (en) | 2020-12-30 | 2020-12-30 | Self-adaptive federal parameter aggregation method for credit assessment of small and micro enterprises |
Publications (1)
Publication Number | Publication Date |
---|---|
CN112634027A true CN112634027A (en) | 2021-04-09 |
Family
ID=75286360
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011600934.0A Pending CN112634027A (en) | 2020-12-30 | 2020-12-30 | Self-adaptive federal parameter aggregation method for credit assessment of small and micro enterprises |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112634027A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113240184A (en) * | 2021-05-21 | 2021-08-10 | 浙江大学 | Building space unit cold load prediction method and system based on federal learning |
CN113469370A (en) * | 2021-06-22 | 2021-10-01 | 河北工业大学 | Industrial Internet of things data sharing method based on federal incremental learning |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111125779A (en) * | 2019-12-17 | 2020-05-08 | 山东浪潮人工智能研究院有限公司 | Block chain-based federal learning method and device |
CN111598143A (en) * | 2020-04-27 | 2020-08-28 | 浙江工业大学 | Credit evaluation-based defense method for federal learning poisoning attack |
CN111600707A (en) * | 2020-05-15 | 2020-08-28 | 华南师范大学 | Decentralized federal machine learning method under privacy protection |
-
2020
- 2020-12-30 CN CN202011600934.0A patent/CN112634027A/en active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111125779A (en) * | 2019-12-17 | 2020-05-08 | 山东浪潮人工智能研究院有限公司 | Block chain-based federal learning method and device |
CN111598143A (en) * | 2020-04-27 | 2020-08-28 | 浙江工业大学 | Credit evaluation-based defense method for federal learning poisoning attack |
CN111600707A (en) * | 2020-05-15 | 2020-08-28 | 华南师范大学 | Decentralized federal machine learning method under privacy protection |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113240184A (en) * | 2021-05-21 | 2021-08-10 | 浙江大学 | Building space unit cold load prediction method and system based on federal learning |
CN113240184B (en) * | 2021-05-21 | 2022-06-24 | 浙江大学 | Building space unit cold load prediction method and system based on federal learning |
CN113469370A (en) * | 2021-06-22 | 2021-10-01 | 河北工业大学 | Industrial Internet of things data sharing method based on federal incremental learning |
CN113469370B (en) * | 2021-06-22 | 2022-08-30 | 河北工业大学 | Industrial Internet of things data sharing method based on federal incremental learning |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11816183B2 (en) | Methods and systems for mining minority-class data samples for training a neural network | |
Dong et al. | Automatic age estimation based on deep learning algorithm | |
CN105740401B (en) | A kind of interested site recommended method and device based on individual behavior and group interest | |
CN108647736B (en) | Image classification method based on perception loss and matching attention mechanism | |
CN107704877A (en) | A kind of image privacy cognitive method based on deep learning | |
CN113657561B (en) | Semi-supervised night image classification method based on multi-task decoupling learning | |
CN110852447A (en) | Meta learning method and apparatus, initialization method, computing device, and storage medium | |
CN112634027A (en) | Self-adaptive federal parameter aggregation method for credit assessment of small and micro enterprises | |
CN108108808A (en) | A kind of position predicting method and device based on depth belief network | |
CN113190688A (en) | Complex network link prediction method and system based on logical reasoning and graph convolution | |
CN109492075B (en) | Transfer learning sequencing method based on loop generation countermeasure network | |
CN108647691A (en) | A kind of image classification method based on click feature prediction | |
CN108960304A (en) | A kind of deep learning detection method of network trading fraud | |
CN109670927A (en) | The method of adjustment and its device of credit line, equipment, storage medium | |
CN112365007B (en) | Model parameter determining method, device, equipment and storage medium | |
Sandeep Kumar et al. | Location identification for real estate investment using data analytics | |
CN111079930B (en) | Data set quality parameter determining method and device and electronic equipment | |
CN109271957A (en) | Face gender identification method and device | |
CN107451617A (en) | One kind figure transduction semisupervised classification method | |
CN112487305B (en) | GCN-based dynamic social user alignment method | |
CN106897388A (en) | Predict the method and device of microblogging event temperature | |
CN115797041A (en) | Financial credit assessment method based on depth map semi-supervised learning | |
CN103544503B (en) | A kind of Activity recognition method based on many examples Markov model | |
CN107402984B (en) | A kind of classification method and device based on theme | |
CN111353525A (en) | Modeling and missing value filling method for unbalanced incomplete data set |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20210409 |
|
RJ01 | Rejection of invention patent application after publication |