ArticlePublisher preview available

BHRAM: a knowledge graph embedding model based on bidirectional and heterogeneous relational attention mechanism

Authors:
To read the full-text of this research, you can request a copy directly from the authors.

Abstract and Figures

Knowledge graph embedding (KGE) is a method designed to predict missing relations between entities in a knowledge graph (KG), which has garnered much attention in recent years due to the incompleteness of KGs. However, existing KGE models have limitations in dealing with heterogeneous KGs and relation direction prediction. To address this issue, a novel KGE model called BHRAM is proposed. It is based on a bidirectional and heterogeneous relational attention mechanism. Specifically, BHRAM comprises three primary components, namely entity aggregation, relation aggregation and triplet prediction. The entity aggregation module divides the adjacency matrix into original and reverse relation adjacency matrices, using graph convolution to aggregate node features and subsequently form entity embedding representations. The relation aggregation module leverages bidirectional relations for feature extraction, learns the weight information of diverse paths independently and generates embedding representations of relation paths through an aggregation function. Finally, the triplet prediction module utilizes a score function for probabilistic predictions. To validate the superiority of BHRAM, comprehensive experiments were conducted on four well-known datasets, including baseline comparisons, relation classification tasks and ablation study. The results demonstrate that BHRAM significantly outperforms the other baselines on the FB15k-237, Kinship and UMLS datasets, while achieving similar or better performance than the baselines on the WN18RR dataset. These findings indicate that BHRAM can serve as a robust and effective model for addressing the heterogeneity in KGs.
This content is subject to copyright. Terms and conditions apply.
Applied Intelligence (2025) 55:245
https://doi.org/10.1007/s10489-024-06212-4
BHRAM: a knowledge graph embedding model based on bidirectional
and heterogeneous relational attention mechanism
Chaoqun Zhang1,2 ·Wanqiu Li1·Yuanbin Mo1·Weidong Tang1·Haoran Li1·Zhilin Zeng1
Accepted: 15 December 2024 / Published online: 30 December 2024
© The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2024
Abstract
Knowledge graph embedding (KGE) is a method designed to predict missing relations between entities in a knowledge graph
(KG), which has garnered much attention in recent years due to the incompleteness of KGs. However, existing KGE models
have limitations in dealing with heterogeneous KGs and relation direction prediction. To address this issue, a novel KGE model
called BHRAM is proposed. It is based on a bidirectional and heterogeneous relational attention mechanism. Specifically,
BHRAM comprises three primary components, namely entity aggregation, relation aggregation and triplet prediction. The
entity aggregation module divides the adjacency matrix into original and reverse relation adjacency matrices, using graph
convolution to aggregate node features and subsequently form entity embedding representations. The relation aggregation
module leverages bidirectional relations for feature extraction, learns the weight information of diverse paths independently and
generates embedding representations of relation paths through an aggregation function. Finally, the triplet prediction module
utilizes a score function for probabilistic predictions. To validate the superiority of BHRAM, comprehensive experiments
were conducted on four well-known datasets, including baseline comparisons, relation classification tasks and ablation study.
The results demonstrate that BHRAM significantly outperforms the other baselines on the FB15k-237, Kinship and UMLS
datasets, while achieving similar or better performance than the baselines on the WN18RR dataset. These findings indicate
that BHRAM can serve as a robust and effective model for addressing the heterogeneity in KGs.
Keywords Attention mechanism ·Graph convolutional network ·Heterogeneity problem ·Knowledge graph embedding
1 Introduction
Knowledge graph (KG) is a structured semantic knowledge
base [1], initially proposed by Google in 2012 [2]. It has
become a crucial component of artificial intelligence (AI) and
has witnessed the emergence of similar KGs such as WordNet
[3], Dbpedia [4], YAGO [5], Freebase [6] and Wikidata [7].
These KGs have significantly contributed to various intelli-
gence driven applications, including recommender systems
[8], information retrieval [9] and question & answer systems
[10].
BWanqiu Li
lwq20211225@163.com
1College of Artificial Intelligence, Guangxi Minzu University,
Nanning 530006, China
2Guangxi Key Laboratory of Hybrid Computation and IC
Design Analysis, Nanning 530006, China
A KG represents concepts and their interrelations in the
physical world using a symbolic form. It stores knowledge
information in the form of directed graphs, as depicted in
Fig. 1. KGs typically consist of nodes that represent entities
and edges that represent relations between neighboring enti-
ties. This representation is commonly expressed as a triplet
(h,r,t), where hdenotes a head entity, rdenotes a relation
and tdenotes a tail entity. For example, (Christopher Nolan,
directed, Batman Begins) indicates that Christopher Nolan
directed the film Batman Begins.
Despite the vast scale of existing KGs, their inherent
incompleteness persists due to the continuous expansion of
real-world knowledge. Knowledge graph embedding (KGE)
has emerged as a promising solution, embedding entities and
relations into continuous vector spaces to preserve struc-
tural information and predict missing facts [11]. The field
of KGE has witnessed the emergence of various methods,
including translational distance-based methods [12], seman-
tic matching-based methods [13] and neural networks-based
123
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
ResearchGate has not been able to resolve any citations for this publication.
Preprint
Full-text available
This paper presents translation-based knowledge graph completion method via efficient relation rotation (TransERR), a straightforward yet effective alternative to traditional translation-based knowledge graph completion models. Different from the previous translation-based models, TransERR encodes knowledge graphs in the hypercomplex-valued space, thus enabling it to possess a higher degree of translation freedom in mining latent information between the head and tail entities. To further minimize the translation distance, TransERR adaptively rotates the head entity and the tail entity with their corresponding unit quaternions, which are learnable in model training. The experiments on 7 benchmark datasets validate the effectiveness and the generalization of TransERR. The results also indicate that TransERR can better encode large-scale datasets with fewer parameters than the previous translation-based models. Our code is available at: \url{https://github.com/dellixx/TransERR}.
Article
The `pre-train, prompt, predict' paradigm of large language models (LLMs) has achieved remarkable success in open-domain question answering (OD-QA). However, few works explore this paradigm in multi-document question answering (MD-QA), a task demanding a thorough understanding of the logical associations among the contents and structures of documents. To fill this crucial gap, we propose a Knowledge Graph Prompting (KGP) method to formulate the right context in prompting LLMs for MD-QA, which consists of a graph construction module and a graph traversal module. For graph construction, we create a knowledge graph (KG) over multiple documents with nodes symbolizing passages or document structures (e.g., pages/tables), and edges denoting the semantic/lexical similarity between passages or document structural relations. For graph traversal, we design an LLM-based graph traversal agent that navigates across nodes and gathers supporting passages assisting LLMs in MD-QA. The constructed graph serves as the global ruler that regulates the transitional space among passages and reduces retrieval latency. Concurrently, the graph traversal agent acts as a local navigator that gathers pertinent context to progressively approach the question and guarantee retrieval quality. Extensive experiments underscore the efficacy of KGP for MD-QA, signifying the potential of leveraging graphs in enhancing the prompt design and retrieval augmented generation for LLMs. Our code: https://github.com/YuWVandy/KG-LLM-MDQA.
Article
The rapid proliferation of online fake news has caused confusion and social panic, underscoring the urgency of effective detection. Although tremendous research effort has been devoted to extracting semantic information from news pieces and detecting fake news regarding their linguistic characteristics, the abundant factual information in Knowledge graphs (KGs) has yet to be explored. There remain two significant challenges to utilizing KGs for fake news detection: 1) The unknown relation between existing knowledge and news content, and 2) fusing characters of fake news from both KGs and news content inherently requires balancing the contributions of these resources, but no prior domain knowledge is available. Therefore, we propose KEHGNN-FD, a novel KG-enhanced Heterogeneous Graph Neural Network, to unleash the power of KGs for fake news detection. We model news content, topics, and entities as a heterogeneous graph and use graph attention to learn high-level news representations by adaptively aggregating neighboring information from the heterogeneous graph and ground-truth KG entities (i.e., Wikidata). Finally, we train a semi-supervised detector to label news as fake or true. Extensive experiments on four fake news datasets show the superiority of KEHGNN-FD over seven baselines regarding accuracy, precision, recall, F1-score, and ROC. The ablation study further validates the efficacy of the KG for fake news detection as well as KEHGNN-FD’s components.
Article
Recently, many researchers have applied Graph Convolutional Neural Networks (GCN) to multi-label learning tasks by establishing relations among labels. However, these researchers have only used GCN as an auxiliary module and have lacked exploration of object relations, which has limited the inference capability of GCN. Applying GCN to multi-label classification in information systems poses significant challenges to address because information systems are considered as a graph with only nodes and no edges. This article proposes a multi-granularity graph convolutional neural network (MG-GCN) based on object relations to enhance the versatility of GCN and demonstrate its reasoning ability. The MG-GCN efficiently solves complex multi-label classification problems in information systems. Firstly, based on the multi-granularity thinking, the attribute set is divided into several attribute subsets based on granularity. Secondly, a graph of differences representing the relationship between objects is established under each attribute subset. Finally, based on the attention mechanism and multiple difference matrices of objects, MG-GCN is proposed and applied to multi-label classification within information systems. The effectiveness and significance of performance of MG-GCN has been verified through ranking-based indicators and hypothesis tests, respectively.