CN115130663A - Heterogeneous network attribute completion method based on graph neural network and attention mechanism - Google Patents

Heterogeneous network attribute completion method based on graph neural network and attention mechanism Download PDF

Info

Publication number
CN115130663A
CN115130663A CN202211043710.3A CN202211043710A CN115130663A CN 115130663 A CN115130663 A CN 115130663A CN 202211043710 A CN202211043710 A CN 202211043710A CN 115130663 A CN115130663 A CN 115130663A
Authority
CN
China
Prior art keywords
attribute
nodes
node
attributes
target node
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202211043710.3A
Other languages
Chinese (zh)
Other versions
CN115130663B (en
Inventor
于彦伟
王凯
董军宇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ocean University of China
Original Assignee
Ocean University of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ocean University of China filed Critical Ocean University of China
Priority to CN202211043710.3A priority Critical patent/CN115130663B/en
Publication of CN115130663A publication Critical patent/CN115130663A/en
Application granted granted Critical
Publication of CN115130663B publication Critical patent/CN115130663B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/082Learning methods modifying the architecture, e.g. adding, deleting or silencing nodes or connections

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Molecular Biology (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)
  • Devices For Executing Special Programs (AREA)

Abstract

The invention discloses a heterogeneous network attribute completion method based on a graph neural network and an attention mechanism, and belongs to the technical field of data processing. Firstly, capturing nodes similar to target nodes (nodes with missing attributes) from a network by combining cosine similarity between K-Nearest Neighbor and the attributes, and converting the attributes of the nodes into network representation of a characteristic domain of the target nodes in a self-adaptive manner; then, based on the attention mechanism of a graph neural network and a transformer, carrying out hierarchical analysis on the topological structure in the network and the attribute information of the nodes, and further obtaining the network representation of the target node in a spatial domain; and finally, model parameter learning is carried out by combining network characterization of the space domain and the characteristic domain based on the Euclidean distance loss function so as to complete the missing attribute. Through practical verification, the attribute completion method provided by the invention has the characteristics of high efficiency and high accuracy.

Description

Heterogeneous network attribute completion method based on graph neural network and attention mechanism
Technical Field
The invention relates to a heterogeneous network attribute completion method based on a graph neural network and an attention mechanism, and belongs to the technical field of data processing.
Background
Networks are ubiquitous in real life, and connections among most objects in the real world can be represented as networks, for example, friend relationships among users can be regarded as social networks, reference relationships among papers can be regarded as reference networks, and connection relationships among road sections can be regarded as a traffic network. The networks described above are all made up of nodes of the same type, and they are therefore also referred to as homogeneous networks. In the real world, heterogeneous networks are more widely available, and the nodes constituting the heterogeneous networks are of different types, for example: the shopping network consists of users and commodities, and the academic network consists of authors, the units where the authors are located and papers. Although these networks contain massive data, the missing attributes of the networks (e.g., some users do not want to upload their own age information in shopping networks; and keywords of papers are not filled in completely by authors in academic networks.) also pose a huge challenge to mining the potential value contained in the networks. The efficiency of data mining on the network can be effectively improved by complementing the missing attributes in the network, but the method is relatively complex, and various factors are considered, such as the connection relation between network nodes and the relation between the existing attributes and the missing attributes of the network nodes. How to efficiently and accurately complement the missing data in the network is also receiving more and more attention from the academic and industrial fields.
In the field of attribute completion, a traditional attribute completion method generally analyzes and completes the connection relation of nodes in a network and semantic text information. The above method does not take into account the topology of the entire network. In recent years, the graph neural network method shows high efficiency and accuracy in mining network information, for example, the graph convolution neural network and the graph attention network show excellent performance in capturing network topology and node attribute information. The development of the graph neural network also brings new possibilities to the field of attribute completion, such as a method for performing joint learning based on the graph convolution neural network in combination with attribute completion and commodity recommendation, and performing attribute completion by using the characteristics learned by the graph neural network. These methods have achieved significant results in attribute completion, but there is still room for improvement.
By analyzing and summarizing the existing attribute completion method, the existing method has the following defects: 1) the network characterization in the heterogeneous network is different from the type of all nodes in the homogeneous network and is the same, and the missing attribute of the nodes in the heterogeneous network needs to be supplemented by comprehensively considering homogeneous node information and heterogeneous node information. This problem has plagued many network characterization models. 2) Graph neural networks do not efficiently capture higher-order nodal information in the network. GCN essentially generates tokens by aggregating the surrounding node information of target nodes under a semi-supervised framework. In the attribute completion problem, the attribute of the target node may be related to not only its surrounding nodes but also its higher-order nodes. If the high-order information of the target node can be captured admittedly only by stacking the number of GCN layers, but as the number of GCN layers increases, the nodes in the network are more similar to the nodes in terms of representation, the specificity is lost, and the accuracy of attribute completion is further influenced.
Disclosure of Invention
In order to solve the problem of heterogeneous network attribute completion more effectively, the invention aims to provide an attribute completion method based on a graph neural network and an attention mechanism, so as to further improve the efficiency and accuracy of attribute completion and provide method and technical support for the heterogeneous network attribute completion problem.
In order to achieve the purpose of the invention, the specific technical scheme adopted by the invention is as follows:
an attribute completion method based on a graph neural network and an attention mechanism comprises the following steps:
s1: acquiring a heterogeneous attribute network with missing attributes, wherein nodes with the missing attributes are called target nodes, and nodes with complete attributes are called source nodes;
s2: selecting K source nodes most similar to the existing attributes of the target nodes by adopting a K-nearest neighbor (K-nearest neighbor) algorithm to perform attribute completion in a feature space;
s3: after K source nodes most similar to the target node are screened out by utilizing cosine similarity, a learnable parameter is given to each source node to dynamically adjust the weight of the attribute of each source node for completing the attribute of the target node, and the attribute representation of the feature space, which is given with the weight, is obtained;
s4: aggregating the nodes directly connected with the target node in the heterogeneous attribute network to obtain the low-order representation of the structure space; the specific aggregation mode is realized by a simplified graph neural network (simplified graph neural networks);
s5: firstly, obtaining a high-order node of a target node in a random walk mode;
s6: different weights are given to nodes in the high-order node sequence based on a transformer, and high-order representation of the structure space, which is given with the weights, is obtained;
s7: splicing the attribute representation of the feature space given with the weight, the low-order representation of the structure space and the high-order representation of the structure space given with the weight, then sending the spliced representations into a multilayer perceptron, and converting the representations into the attributes of the target nodes;
s8: and (3) adopting a supervised learning mode, firstly artificially removing partial attributes of partial nodes, then complementing the attributes through reconstruction prediction, continuously training through the difference between the complemented attributes and the real attributes, and finally complementing the attribute missing values of other nodes by using a trained model.
Further, in S1, a heterogeneous network with missing attributes is defined as
Figure 185024DEST_PATH_IMAGE001
A collection of vertices in the diagram is represented,
Figure 164482DEST_PATH_IMAGE002
a collection of edges in the diagram is represented,
Figure 6536DEST_PATH_IMAGE003
representing a matrix of attributes of the vertices in the graph,
Figure 464062DEST_PATH_IMAGE004
is the number of vertices in the graph,
Figure 544013DEST_PATH_IMAGE005
for the dimensions of the features of each vertex,
Figure 428793DEST_PATH_IMAGE006
is a matrix of marks when
Figure 23722DEST_PATH_IMAGE007
When the utility model is used, the water is discharged,
Figure 284939DEST_PATH_IMAGE008
is missing; on the contrary, the first step is to take the reverse,
Figure 16135DEST_PATH_IMAGE009
the corresponding attribute in (a) is complete.
Further, in the S2, a large number of nodes are usually included in the heterogeneous network, and it is inefficient and impractical to supplement the missing attribute of the target node with the attributes of all the source nodes; therefore, K neighbor is adopted to select K source nodes most similar to the existing attributes of the target node to complete the attributes in the feature space, and cosine similarity (formula (1)) is used to measure the similarity of two attribute vectors;
Figure 71816DEST_PATH_IMAGE010
(1)
wherein ,
Figure 888462DEST_PATH_IMAGE011
which indicates the similarity of the two nodes,
Figure 687791DEST_PATH_IMAGE012
representing an existing property of the target node,
Figure 804651DEST_PATH_IMAGE013
representing attributes in the source node that correspond to existing attributes of the target node,
Figure 296812DEST_PATH_IMAGE014
a larger value indicates a higher similarity between the target node and the source node.
Further, in the S3, there may be a plurality of factors affecting the attribute completion relationship of the source node to the target node in the heterogeneous network, for example: whether nodes are directly connected with each other, the weight of edges of the connected nodes, and the like. The dynamic adjustment is performed according to equation (2):
Figure 600755DEST_PATH_IMAGE015
(2)
wherein ,
Figure 938195DEST_PATH_IMAGE016
for the characterization of the feature space of the target node,
Figure 643983DEST_PATH_IMAGE017
is with the target node
Figure 244729DEST_PATH_IMAGE018
Most similar of
Figure 832705DEST_PATH_IMAGE019
A set of a plurality of source nodes, each source node,
Figure 973836DEST_PATH_IMAGE020
for each learnable adjustment weight corresponding to a source node,
Figure 534131DEST_PATH_IMAGE021
is the feature vector of the source node.
Further, in S4, after completing the feature learning of the target node in the feature space, learning a feature of the target node in the structural space is required; the specific formula (3) is shown in the following:
Figure 571357DEST_PATH_IMAGE022
(3)
wherein ,
Figure 584312DEST_PATH_IMAGE023
for simplification of the figure in neural networks
Figure 325872DEST_PATH_IMAGE024
The weight matrix of the layer or layers is,
Figure 740673DEST_PATH_IMAGE025
is the first
Figure 745538DEST_PATH_IMAGE024
Output of layer at the first layer of the reduced graph neural network
Figure 980210DEST_PATH_IMAGE026
Is a contiguous matrix.
The graph neural network essentially captures topology structure information and attribute information of nodes in the network by aggregating neighbor attributes of target nodes, but with stacking of the layers of the graph neural network, the characteristics aggregated by the target nodes are over-smoothed, that is, the characteristics obtained by the target nodes through the aggregated characteristics lose specificity and distinctiveness. Therefore, the neighbor nodes of the target node are aggregated in the structural space by using the simplified graph neural network, and only one layer of network is used, so that the characteristics of the topology and the attributes are captured while the specificity of the target node representation is kept.
Further, the step S5 is specifically: in order to capture the high-order information of a target node in the heterogeneous network, the high-order node of the target node is obtained by using a random walk mode, wherein the random walk mode is a mode of traversing nodes in a graph by combining the advantages of depth-first traversal and breadth-first traversal in the graph, and the specific traversal mode is shown as a formula (4):
Figure 259882DEST_PATH_IMAGE027
(4)
wherein
Figure 325927DEST_PATH_IMAGE028
Representing nodes
Figure 236114DEST_PATH_IMAGE029
To the node
Figure 223661DEST_PATH_IMAGE030
The weight of the edge of (c);
and for each target node, taking the target node as a root node, and performing random walk based on the target node to obtain a node sequence. Since directly connected source nodes are already used in the reduced graph neural network, these nodes are deleted in the randomly walked node sequence.
Further, in S6, since the influence of the node sequence generated by the random walk on the target node is different, the nodes in the node sequence are given different weights based on the transform. The method specifically comprises the following steps:
s6-1: firstly, carrying out linear change on the characteristics of sequences of a target node and a source node, then calculating the weights of the target node and the nodes in each sequence based on the linear change, and carrying out softmax normalization operation on the obtained weights for the calculation stability;
s6-2: then, a new linear transformation independent of the weight calculation is performed again on the node features in the source node sequence.
S6-3: and finally, endowing each linear transformed source node sequence node feature with weight, and accumulating the linear transformed source node sequence node features to obtain a high-order node representation of a target node, wherein the calculation method is shown as a formula (5):
Figure 307024DEST_PATH_IMAGE031
(5)
wherein
Figure 165259DEST_PATH_IMAGE032
Representing nodes
Figure 43085DEST_PATH_IMAGE033
By simplifying the characterization that the neural network of the graph aggregates,
Figure 314666DEST_PATH_IMAGE034
representing nodes
Figure 428419DEST_PATH_IMAGE035
Is characterized in that it is a mixture of two or more of the above-mentioned components,
Figure 406740DEST_PATH_IMAGE036
Figure 455467DEST_PATH_IMAGE037
and
Figure 417607DEST_PATH_IMAGE038
a projection parameter matrix representing the parameters that can be learned,
Figure 577193DEST_PATH_IMAGE039
representing nodes
Figure 206757DEST_PATH_IMAGE035
High-order network characterization of (1).
S6-4: on the basis, the attention mechanism is expanded to a multi-head attention mechanism to capture multiple dependency relationships between the target node and the high-order source node. Then, the multiple dependency relationships are sent to an average pooling layer to obtain a final high-order network characterization, and the calculation method is shown as formula (6):
Figure 629648DEST_PATH_IMAGE040
(6)
wherein ,
Figure 875822DEST_PATH_IMAGE041
representing nodes
Figure 839099DEST_PATH_IMAGE042
In the first place
Figure 995274DEST_PATH_IMAGE043
The resulting high-order network characterization in the secondary attention mechanism,
Figure 854645DEST_PATH_IMAGE044
indicating the number of times a total attention mechanism calculation is required.
Further, in S7, the specific calculation manner of the splicing is shown in formula (7):
Figure 322536DEST_PATH_IMAGE045
(7)
wherein
Figure 761607DEST_PATH_IMAGE046
A vector splicing operation is represented as a vector splicing operation,
Figure 834605DEST_PATH_IMAGE047
for predicted target node
Figure 864878DEST_PATH_IMAGE048
The attribute value of (2).
Further, the S8 specifically includes:
s8-1: after the predicted attribute of the target node is obtained, the existing attribute is reserved, the predicted attribute is filled into the missing attribute to complete the task of attribute completion, and the filling method is shown as formula (8):
Figure 23327DEST_PATH_IMAGE049
(8)
wherein ,
Figure 62827DEST_PATH_IMAGE050
representing the predicted attributes of all nodes after attribute completion,
Figure 724753DEST_PATH_IMAGE051
the product of the hadamard is represented,
Figure 191506DEST_PATH_IMAGE052
is a matrix with elements all 1:
s8-2: setting a gap between the loss function measurement prediction attribute and the real attribute based on an Euclidean distance formula, wherein a specific calculation mode is shown as a formula (9):
Figure 837251DEST_PATH_IMAGE053
(9)
wherein
Figure 883705DEST_PATH_IMAGE054
A set of target nodes is represented that,
Figure 931295DEST_PATH_IMAGE055
is a node
Figure 303371DEST_PATH_IMAGE042
The true value of the missing attribute(s),
Figure 436412DEST_PATH_IMAGE056
is a node
Figure 552135DEST_PATH_IMAGE035
Predicted missing attribute values.
The invention considers the relevance between the target node and the source node in the attribute space and the relevance between the target node and the source node in the structure space when the heterogeneous network attribute is complemented, which is specifically expressed as follows: in attribute space use
Figure 454232DEST_PATH_IMAGE057
Neighbor algorithm finds the most similar to target node
Figure 997209DEST_PATH_IMAGE057
A source node then adaptively assigns to it a corresponding
Figure 351967DEST_PATH_IMAGE057
And adding the weight values to obtain attribute space representation of the target node. In a structural space, aggregating first-order neighbor information of a target node by using a simplified graph neural network to obtain a low-order representation of the target node; and then capturing high-order neighbor information of the target node in the structural space by using a multi-head attention mechanism based on transformer and random walk to obtain high-order representation of the target node. Finally, the three characteristics are fused, and the loss function is introduced based on Euclidean distanceAnd (5) leading to updating parameters of the integral model, and finally completing the missing attribute of the target node.
The invention has the advantages and beneficial effects that:
compared with the traditional attribute completion method, the method introduces a network representation learning framework to perform attribute completion, and can better capture the structural characteristics of the target node in the network; the first-order neighbors of the target node are aggregated by using the simplified graph neural network, so that the accuracy of the obtained network representation can be ensured on the premise of improving the attribute completion time efficiency; the method uses an attention mechanism based on random walk and transformer to capture high-order neighbor information of a target node, and solves the problem of node representation over-smoothness caused by a simple stacking graph neural network module to a certain extent.
Through practical verification, the attribute completion method provided by the invention has the characteristics of high efficiency and high accuracy.
Drawings
FIG. 1 is an overall flow chart of the present invention.
Fig. 2 is a framework diagram of the present invention.
Fig. 3 is a topology structural diagram one of the weights between nodes obtained in the present invention.
Fig. 4 is a topology structural diagram two of the weights between the nodes obtained in the present invention.
FIG. 5 is a flow chart of the present invention for obtaining high-order node characterization based on an attention mechanism.
Detailed Description
The invention will be further described with reference to the accompanying figures 1-5 and the specific embodiments.
Example 1:
an attribute completion method based on a graph neural network and an attention mechanism is shown in the overall flow chart of fig. 1. The method comprises the following steps:
s1: acquiring a heterogeneous attribute network with missing attributes, wherein nodes with the missing attributes are called target nodes, and nodes with complete attributes are called source nodes; defining a heterogeneous attribute network with missing attributes as
Figure 271381DEST_PATH_IMAGE058
Representing a collection of vertices in a graph
Figure 27985DEST_PATH_IMAGE059
Figure 741863DEST_PATH_IMAGE002
A collection of edges in the diagram is represented,
Figure 115075DEST_PATH_IMAGE003
representing the attribute matrix of the vertexes in the graph, n is the number of the vertexes in the graph,
Figure 572602DEST_PATH_IMAGE060
for the dimensions of the features of each vertex,
Figure 183711DEST_PATH_IMAGE061
is a matrix of marks, if
Figure 334070DEST_PATH_IMAGE062
Then represents
Figure 663420DEST_PATH_IMAGE063
Is missing; otherwise, it means
Figure 924637DEST_PATH_IMAGE064
Is complete;
s2: selecting K source nodes most similar to the existing attributes of the target nodes by adopting a K-nearest neighbor (K-nearest neighbor) algorithm to complete the attributes in the feature space; in heterogeneous networks, which typically include a large number of nodes, it is inefficient and impractical to use the attributes of all source nodes to fill in the missing attributes of the target nodes; therefore, K source nodes most similar to the existing attributes of the target node are selected by adopting K neighbors to perform attribute completion in a feature space, wherein cosine similarity (formula (1)) is used for measuring the similarity of two attribute vectors;
Figure 390254DEST_PATH_IMAGE065
(1)
wherein ,
Figure 711514DEST_PATH_IMAGE066
which indicates the similarity of the two nodes,
Figure 528160DEST_PATH_IMAGE067
representing an existing property of the target node,
Figure 327489DEST_PATH_IMAGE068
representing attributes in the source node that correspond to existing attributes of the target node,
Figure 913191DEST_PATH_IMAGE069
the larger the value is, the higher the similarity between the target node and the source node is;
s3: after K source nodes most similar to the target node are screened out by utilizing cosine similarity, a learnable parameter is endowed to each source node to dynamically adjust the weight of the attribute of each source node for completing the attribute of the target node, and attribute representation endowed with the weight of a feature space is obtained; in a heterogeneous network, there may be a plurality of factors that affect the attribute completion relationship of the source node to the target node, for example: whether nodes are directly connected with each other, the weight of edges of the connected nodes, and the like. The dynamic adjustment is performed according to equation (2):
Figure 405352DEST_PATH_IMAGE070
(2)
wherein ,
Figure 443715DEST_PATH_IMAGE071
for the characterization of the feature space of the target node,
Figure 46735DEST_PATH_IMAGE072
is with the target node
Figure 752523DEST_PATH_IMAGE073
Most similar of
Figure 150006DEST_PATH_IMAGE074
A set of a plurality of source nodes, each of the source nodes having a node address,
Figure 941244DEST_PATH_IMAGE075
for each learnable adjustment weight corresponding to a source node,
Figure 82376DEST_PATH_IMAGE076
a feature vector of a source node;
s4: aggregating the nodes directly connected with the target node in the heterogeneous attribute network to obtain the low-order representation of the structure space; the specific aggregation mode is realized by a simplified graph neural network (simplified graph relational networks); after the feature space finishes the feature learning of the target node, the feature of the target node needs to be learned in the structural space; the specific formula (3) is shown in the following:
Figure 642670DEST_PATH_IMAGE077
(3)
wherein ,
Figure 414317DEST_PATH_IMAGE078
for simplification of the figure in neural networks
Figure 692852DEST_PATH_IMAGE079
The weight matrix of the layer(s) is,
Figure 637674DEST_PATH_IMAGE080
is the first
Figure 52475DEST_PATH_IMAGE079
Output of layer at the first layer of the reduced graph neural network
Figure 57340DEST_PATH_IMAGE081
Is a contiguous matrix;
s5: firstly, obtaining a high-order node of a target node in a random walk mode; in order to capture high-order information of a target node in a heterogeneous network, a random walk mode is firstly used for obtaining the high-order node of the target node, the random walk mode is a mode for traversing nodes in a graph by combining the advantages of depth-first traversal and breadth-first traversal in the graph, and the specific traversal mode is shown in a formula (4):
Figure 823170DEST_PATH_IMAGE082
(4)
wherein
Figure 40525DEST_PATH_IMAGE083
Representing nodes
Figure 840991DEST_PATH_IMAGE084
To node
Figure 751178DEST_PATH_IMAGE085
The weight of the edge of (1);
and for each target node, taking the target node as a root node, and performing random walk based on the target node to obtain a node sequence. Because directly connected source nodes are already used in the reduced graph neural network, the nodes are deleted in the randomly walked node sequence;
s6: different weights are given to nodes in the high-order node sequence based on a transformer, and high-order representation of the structure space, which is given with the weights, is obtained; the method comprises the following specific steps:
s6-1: firstly, carrying out linear change on the characteristics of sequences of a target node and a source node, then calculating the weights of the target node and the nodes in each sequence based on the linear change, and carrying out softmax normalization operation on the obtained weights for the calculation stability;
s6-2: then, a new linear transformation independent of the weight calculation is performed again on the node features in the source node sequence.
S6-3: and finally, giving the weight to the node characteristics of each source node sequence after linear transformation, and accumulating the node characteristics to obtain the high-order node characterization of the target node, wherein the calculation method is shown as a formula (5):
Figure 738726DEST_PATH_IMAGE086
(5)
wherein
Figure 25351DEST_PATH_IMAGE087
Representing nodes
Figure 414744DEST_PATH_IMAGE088
By simplifying the characterization that the neural network of the graph aggregates,
Figure 761411DEST_PATH_IMAGE089
representing nodes
Figure 259692DEST_PATH_IMAGE088
Is characterized in that the pressure difference between the pressure sensor and the pressure sensor,
Figure 84429DEST_PATH_IMAGE090
Figure 328329DEST_PATH_IMAGE091
and
Figure 845898DEST_PATH_IMAGE092
a projection parameter matrix representing the parameters that can be learned,
Figure 808037DEST_PATH_IMAGE093
representing nodes
Figure 436465DEST_PATH_IMAGE088
High-order network characterization of (1).
S6-4: on the basis, the attention mechanism is expanded to a multi-head attention mechanism to capture multiple dependency relationships between a target node and a high-order source node. Then, the multiple dependency relationships are sent to an average pooling layer to obtain a final high-order network characterization, and the calculation method is as shown in formula (6):
Figure 534871DEST_PATH_IMAGE094
(6)
wherein ,
Figure 223341DEST_PATH_IMAGE095
representing nodes
Figure 672777DEST_PATH_IMAGE088
In the first place
Figure 42578DEST_PATH_IMAGE096
The resulting high-order network characterization in the secondary attention mechanism,
Figure 261070DEST_PATH_IMAGE097
representing the number of times a total attention mechanism calculation is required;
s7: splicing the attribute representation of the feature space given with the weight, the low-order representation of the structure space and the high-order representation of the structure space given with the weight, then sending the spliced representations into a multilayer perceptron, and converting the representations into the attributes of the target nodes; the specific calculation mode of the splicing is shown as formula (7):
Figure 386021DEST_PATH_IMAGE098
(7)
wherein
Figure 57174DEST_PATH_IMAGE099
A vector splicing operation is represented as a vector splicing operation,
Figure 496245DEST_PATH_IMAGE100
for predicted target node
Figure 569244DEST_PATH_IMAGE088
The attribute value of (2).
S8: adopting a supervised learning mode, firstly manually removing partial attributes of partial nodes, then complementing the attributes through reconstruction prediction, continuously training through the difference between the complemented attributes and the real attributes, and finally complementing attribute missing values of other nodes by using a trained model; the method specifically comprises the following steps:
s8-1: after the predicted attribute of the target node is obtained, the existing attribute is reserved, the predicted attribute is filled into the missing attribute to complete the task of attribute completion, and the filling method is shown as formula (8):
Figure 865096DEST_PATH_IMAGE101
(8)
wherein ,
Figure 23545DEST_PATH_IMAGE102
representing the predicted attributes of all nodes after attribute completion,
Figure 531886DEST_PATH_IMAGE103
the product of the hadamard is represented,
Figure 724970DEST_PATH_IMAGE104
is a matrix with elements all being 1;
s8-2: setting a gap between the loss function measurement prediction attribute and the real attribute based on an Euclidean distance formula, wherein a specific calculation mode is shown as a formula (9):
Figure 926144DEST_PATH_IMAGE105
(9)
wherein
Figure 571889DEST_PATH_IMAGE106
A set of target nodes is represented as,
Figure 883922DEST_PATH_IMAGE107
is a node
Figure 931513DEST_PATH_IMAGE108
The true value of the missing attribute(s),
Figure 506850DEST_PATH_IMAGE109
is a node
Figure 639891DEST_PATH_IMAGE110
Predicted missing attribute values.
Example 2: this example uses example 1 as a basic method to perform module design.
An attribute completion system based on a graph neural network and an attention mechanism comprises a data preprocessing module, a mark matrix construction module, a feature space characterization learning module, a low-order neighbor characterization learning module, a high-order neighbor characterization learning module, a characterization fusion module and an attribute reasoning module, and as shown in fig. 2, the following describes each part in detail:
the data preprocessing module: firstly, normalizing attribute features in an original data set, then dividing the data into a training set, a testing set and a verification set, randomly removing attribute information of nodes in the training set, and recording the removed attribute information as a true value to guide a model to learn.
The mark matrix constructing module: traversing the node attributes in the data set, marking the missing attribute values of the nodes, and further forming an attribute marking matrix
Figure 755615DEST_PATH_IMAGE111
The feature space characterization learning module: and selecting similar nodes of the target node in the structural space, giving weights to the nodes, and summing the characteristics of the nodes to obtain the characterization of the attribute space of the target node.
The low-order neighbor representation learning module: attributes of first-order neighbor nodes of the target node are aggregated by using the simplified graph neural network, and as shown in fig. 3, low-order neighbor representations of the target node are obtained.
The high-order neighbor representation learning module: aggregating the attributes of the higher-order neighbor nodes of the target node through random walk and a transformer-based attention mechanism, as shown in fig. 4 and 5, obtaining the higher-order neighbor representation of the target node.
The characterization fusion module: and fusing the feature space representation, the low-order neighbor node representation and the high-order neighbor node representation of the target node.
The attribute reasoning module: and predicting the fusion representation of the target node through a multilayer perceptron to obtain the prediction attribute of the target node, and filling the predicted attribute into the corresponding missing attribute through a marking matrix.
Example 3: the embodiment performs instance verification based on the method and system
In order to verify the accuracy of the model attribute completion proposed by the invention, in three data sets: experiments were conducted on DataBase systems and program Logic networks (DBLP), Association for Computing Machinery (ACM), and Internet Movie databases (IMDb), using Heat Kernel and Correlation as evaluation indices, and comparing them with seven existing models.
The seven existing models are respectively: matrix Completion (MC), maximum likelihood Estimation (EM), multi-layer Perceptron (MLP)
) Support Vector Regression (SVR), Heterogeneous Graph attention networks (HGAT), Adaptive Graph Neural networks (AGCN), and Heterogeneous Graph Neural networks by Attribute Completion (HGNN-AC).
TABLE 1 comparative experimental results
Figure 657712DEST_PATH_IMAGE113
The final experimental results are shown in table 1, wherein AC-HEN is the method provided by the present invention. It can be seen that on three real data sets, Heat Kernel and Correlation of the attribute completion method provided by the invention are significantly higher than those of other methods, which means that the model constructed by the invention is superior to other existing models, and the accuracy of attribute completion is higher.
The above-mentioned embodiments are merely intended to be examples of the present invention, but the scope of the present invention is not limited thereto, and all those skilled in the art can understand that the substitutions and changes within the technical scope of the present invention are included in the present invention, so the present invention shall be subject to the protection scope of the claims.

Claims (9)

1. A heterogeneous network attribute completion method based on a graph neural network and an attention mechanism is characterized by comprising the following steps:
s1: acquiring a heterogeneous attribute network with missing attributes, wherein nodes with the missing attributes are called target nodes, and nodes with complete attributes are called source nodes;
s2: selecting K source nodes which are most similar to the existing attributes of the target node by adopting a K neighbor algorithm to perform attribute completion in a feature space;
s3: after K source nodes most similar to the target node are screened out by utilizing cosine similarity, a learned parameter is given to each source node to dynamically adjust the weight of the attribute of each source node for completing the attribute of the target node, and the attribute representation of the feature space, which is given with the weight, is obtained;
s4: aggregating the nodes directly connected with the target node in the heterogeneous attribute network to obtain the low-order representation of the structure space; the specific aggregation mode is realized by a simplified graph neural network;
s5: firstly, obtaining a high-order node of a target node in a random walk mode;
s6: different weights are given to nodes in the high-order node sequence based on a transformer, and high-order representation of the structure space, which is given with the weights, is obtained;
s7: splicing the attribute representation of the feature space given with the weight, the low-order representation of the structure space and the high-order representation of the structure space given with the weight, then sending the spliced representations into a multilayer perceptron, and converting the representations into the attributes of the target nodes;
s8: and (3) adopting a supervised learning mode, firstly manually removing partial attributes of partial nodes, then complementing the attributes through reconstruction prediction, continuously training through the difference between the complemented attributes and the real attributes, and finally complementing the attribute missing values of other nodes by utilizing a trained model.
2. The attribute completion method according to claim 1, wherein in S1, a heterogeneous network with missing attributes is defined as
Figure 511699DEST_PATH_IMAGE001
A collection of vertices in the diagram is represented,
Figure 824869DEST_PATH_IMAGE002
a collection of edges in the diagram is represented,
Figure 196944DEST_PATH_IMAGE003
representing a matrix of attributes of the vertices in the graph,
Figure 329985DEST_PATH_IMAGE004
is the number of vertices in the graph,
Figure 445709DEST_PATH_IMAGE005
for the dimension of each of the vertex features,
Figure 347806DEST_PATH_IMAGE006
is a matrix of marks when
Figure 890783DEST_PATH_IMAGE007
When the temperature of the water is higher than the set temperature,
Figure 245541DEST_PATH_IMAGE008
is missing; on the contrary, the method can be used for carrying out the following steps,
Figure 899376DEST_PATH_IMAGE009
the corresponding attribute in (a) is complete.
3. The method according to claim 1, wherein in S2, K neighbors are first used to select K source nodes that are most similar to the existing attributes of the target node for attribute completion in the feature space, and cosine similarity is used to measure the similarity between two attribute vectors;
Figure 655979DEST_PATH_IMAGE010
(1)
wherein ,
Figure 369857DEST_PATH_IMAGE011
which indicates the similarity of the two nodes,
Figure 477491DEST_PATH_IMAGE012
representing an existing property of the target node,
Figure 935017DEST_PATH_IMAGE013
representing attributes in the source node that correspond to existing attributes of the target node,
Figure 546127DEST_PATH_IMAGE014
a larger value indicates a higher similarity between the target node and the source node.
4. The property completion method according to claim 1, wherein in said S3, said dynamic adjustment is performed according to formula (2):
Figure 696485DEST_PATH_IMAGE015
(2)
wherein ,
Figure 25836DEST_PATH_IMAGE016
for the characterization of the feature space of the target node,
Figure 310490DEST_PATH_IMAGE017
is with the target node
Figure 41686DEST_PATH_IMAGE018
Is most similar to
Figure 362946DEST_PATH_IMAGE019
A set of a plurality of source nodes, each of the source nodes having a node address,
Figure 914013DEST_PATH_IMAGE020
for each learnable adjustment weight corresponding to a source node,
Figure 978921DEST_PATH_IMAGE021
is the feature vector of the source node.
5. The method for attribute completion according to claim 1, wherein in S4, after completing the feature learning of the target node in the feature space, the feature needs to be learned in the structure space; the specific formula (3) is shown in the following:
Figure 299044DEST_PATH_IMAGE022
(3)
wherein ,
Figure 525626DEST_PATH_IMAGE023
for simplifying the first in the neural network
Figure 829568DEST_PATH_IMAGE024
The weight matrix of the layer or layers is,
Figure 698167DEST_PATH_IMAGE025
is the first
Figure 138375DEST_PATH_IMAGE024
Output of layer at the first layer of the reduced graph neural network
Figure 535859DEST_PATH_IMAGE026
Is a contiguous matrix.
6. The attribute completion method according to claim 1, wherein S5 specifically: in order to capture the high-order information of the target node in the heterogeneous network, the high-order node of the target node is obtained by using a random walk mode, wherein the random walk mode is a mode of traversing nodes in a graph by combining the advantages of depth-first traversal and breadth-first traversal, and the specific traversal mode is as shown in formula (4):
Figure 327097DEST_PATH_IMAGE027
(4)
wherein
Figure 733808DEST_PATH_IMAGE028
Representing nodes
Figure 294102DEST_PATH_IMAGE029
To the node
Figure 862487DEST_PATH_IMAGE030
The weight of the edge of (2).
7. The attribute completion method according to claim 1, wherein the S6 is specifically:
s6-1: firstly, carrying out linear change on the characteristics of a target node sequence and a source node sequence, then calculating the weights of the target node and the nodes in each sequence based on the linear change, and carrying out softmax normalization operation on the obtained weights for the calculation stability;
s6-2: then, carrying out new linear transformation independent of weight calculation on the node characteristics in the source node sequence again;
s6-3: finally, giving weight to the node characteristics of each source node sequence after linear transformation, and accumulating the node characteristics to obtain the high-order node representation of the target node; the calculation method is shown in formula (5):
Figure 141021DEST_PATH_IMAGE031
(5)
wherein
Figure 85844DEST_PATH_IMAGE032
Representing nodes
Figure 500644DEST_PATH_IMAGE033
By simplifying the characterization that the neural network of the graph aggregates,
Figure 239930DEST_PATH_IMAGE034
representing nodes
Figure 740182DEST_PATH_IMAGE035
Is characterized in that it is a mixture of two or more of the above-mentioned components,
Figure 488695DEST_PATH_IMAGE036
Figure 758002DEST_PATH_IMAGE037
and
Figure 933769DEST_PATH_IMAGE038
a projection parameter matrix representing the parameters that can be learned,
Figure 921316DEST_PATH_IMAGE039
representing nodes
Figure 207941DEST_PATH_IMAGE035
High-order network characterization of (1);
s6-4: expanding the attention mechanism to a multi-head attention mechanism to capture multiple dependency relationships between a target node and a high-order source node; then, the multiple dependency relationships are sent to an average pooling layer to obtain a final high-order network characterization, and the calculation method is shown as formula (6):
Figure 597334DEST_PATH_IMAGE040
(6)
wherein ,
Figure 944002DEST_PATH_IMAGE041
representing nodes
Figure 153266DEST_PATH_IMAGE033
In the first place
Figure 243582DEST_PATH_IMAGE042
The resulting high-order network characterization in the secondary attention mechanism,
Figure 487482DEST_PATH_IMAGE043
indicating the number of times a total attention mechanism calculation is required.
8. The attribute completion method according to claim 1, wherein in S7, the specific calculation manner of the concatenation is as shown in formula (7):
Figure 5051DEST_PATH_IMAGE044
(7)
wherein
Figure 701611DEST_PATH_IMAGE045
A vector splicing operation is represented as a function of,
Figure 330039DEST_PATH_IMAGE046
for predicted target node
Figure 428445DEST_PATH_IMAGE047
The attribute value of (2).
9. The attribute completion method according to claim 1, wherein the S8 is specifically:
s8-1: after the predicted attribute of the target node is obtained, the existing attribute is reserved, the predicted attribute is filled into the missing attribute to complete the task of attribute completion, and the filling method is shown as formula (8):
Figure 116915DEST_PATH_IMAGE048
(8)
wherein ,
Figure 300772DEST_PATH_IMAGE049
representing the predicted attributes of all nodes after attribute completion,
Figure 732890DEST_PATH_IMAGE050
the product of the hadamard is represented,
Figure 216961DEST_PATH_IMAGE051
is a matrix with elements all being 1;
s8-2: setting a gap between the loss function measurement prediction attribute and the real attribute based on an Euclidean distance formula, wherein a specific calculation mode is shown as a formula (9):
Figure 76332DEST_PATH_IMAGE052
(9)
wherein
Figure 481906DEST_PATH_IMAGE053
A set of target nodes is represented that,
Figure 717715DEST_PATH_IMAGE054
is a node
Figure 525134DEST_PATH_IMAGE033
The true value of the missing attribute(s),
Figure 555407DEST_PATH_IMAGE055
is a node
Figure 713856DEST_PATH_IMAGE035
Predicted missing attribute values.
CN202211043710.3A 2022-08-30 2022-08-30 Heterogeneous network attribute completion method based on graph neural network and attention mechanism Active CN115130663B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211043710.3A CN115130663B (en) 2022-08-30 2022-08-30 Heterogeneous network attribute completion method based on graph neural network and attention mechanism

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211043710.3A CN115130663B (en) 2022-08-30 2022-08-30 Heterogeneous network attribute completion method based on graph neural network and attention mechanism

Publications (2)

Publication Number Publication Date
CN115130663A true CN115130663A (en) 2022-09-30
CN115130663B CN115130663B (en) 2023-10-13

Family

ID=83388013

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211043710.3A Active CN115130663B (en) 2022-08-30 2022-08-30 Heterogeneous network attribute completion method based on graph neural network and attention mechanism

Country Status (1)

Country Link
CN (1) CN115130663B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115759199A (en) * 2022-11-21 2023-03-07 山东大学 Multi-robot environment exploration method and system based on hierarchical graph neural network

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113094484A (en) * 2021-04-07 2021-07-09 西北工业大学 Text visual question-answering implementation method based on heterogeneous graph neural network
WO2021179640A1 (en) * 2020-03-10 2021-09-16 深圳大学 Graph model-based short video recommendation method, intelligent terminal and storage medium
WO2021179838A1 (en) * 2020-03-10 2021-09-16 支付宝(杭州)信息技术有限公司 Prediction method and system based on heterogeneous graph neural network model
CN114692867A (en) * 2022-03-24 2022-07-01 大连理工大学 Network representation learning algorithm combining high-order structure and attention mechanism
CN114723037A (en) * 2022-02-25 2022-07-08 上海理工大学 Heterogeneous graph neural network computing method for aggregating high-order neighbor nodes

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021179640A1 (en) * 2020-03-10 2021-09-16 深圳大学 Graph model-based short video recommendation method, intelligent terminal and storage medium
WO2021179838A1 (en) * 2020-03-10 2021-09-16 支付宝(杭州)信息技术有限公司 Prediction method and system based on heterogeneous graph neural network model
CN113094484A (en) * 2021-04-07 2021-07-09 西北工业大学 Text visual question-answering implementation method based on heterogeneous graph neural network
CN114723037A (en) * 2022-02-25 2022-07-08 上海理工大学 Heterogeneous graph neural network computing method for aggregating high-order neighbor nodes
CN114692867A (en) * 2022-03-24 2022-07-01 大连理工大学 Network representation learning algorithm combining high-order structure and attention mechanism

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
HONGYAN CUI等: ""Self-training method based on GCN for semi-supervised short text classification"", 《INFORMATION SCIENCES》 *
丁钰;魏浩;潘志松;刘鑫;: "网络表示学习算法综述", 计算机科学, no. 09 *
杨宝生;: "基于注意力机制增强图卷积神经网络的个性化新闻推荐", 兰州文理学院学报(自然科学版), no. 05 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115759199A (en) * 2022-11-21 2023-03-07 山东大学 Multi-robot environment exploration method and system based on hierarchical graph neural network
CN115759199B (en) * 2022-11-21 2023-09-26 山东大学 Multi-robot environment exploration method and system based on hierarchical graph neural network

Also Published As

Publication number Publication date
CN115130663B (en) 2023-10-13

Similar Documents

Publication Publication Date Title
CN112529168B (en) GCN-based attribute multilayer network representation learning method
CN109389151B (en) Knowledge graph processing method and device based on semi-supervised embedded representation model
CN113961759B (en) Abnormality detection method based on attribute map representation learning
CN113255895B (en) Structure diagram alignment method and multi-diagram joint data mining method based on diagram neural network representation learning
CN114565053B (en) Deep heterogeneous graph embedded model based on feature fusion
Choi et al. Identifying emerging technologies to envision a future innovation ecosystem: A machine learning approach to patent data
CN115859793A (en) Attention-based method and system for detecting abnormal behaviors of heterogeneous information network users
CN112784118A (en) Community discovery method and device in graph sensitive to triangle structure
CN115130663B (en) Heterogeneous network attribute completion method based on graph neural network and attention mechanism
Ma et al. Class-imbalanced learning on graphs: A survey
CN115699058A (en) Feature interaction through edge search
Du et al. Image recommendation algorithm combined with deep neural network designed for social networks
CN114936307A (en) Method for constructing normal graph model
CN115114484A (en) Abnormal event detection method and device, computer equipment and storage medium
Sheng et al. Personalized recommendation of location-based services using spatio-temporal-aware long and short term neural network
CN117408336A (en) Entity alignment method for structure and attribute attention mechanism
CN117093928A (en) Self-adaptive graph node anomaly detection method based on spectral domain graph neural network
CN115545833A (en) Recommendation method and system based on user social information
Tang et al. Hypergraph structure inference from data under smoothness prior
Zhou et al. A structure distinguishable graph attention network for knowledge base completion
CN113744023A (en) Dual-channel collaborative filtering recommendation method based on graph convolution network
Chen et al. Semi-supervised heterogeneous graph learning with multi-level data augmentation
CN116702214B (en) Privacy data release method and system based on coherent proximity and Bayesian network
CN116402589B (en) Commodity recommendation method and system based on knowledge graph and rotary coding
CN114936296B (en) Indexing method, system and computer equipment for super-large-scale knowledge map storage

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant