Next Article in Journal
Motion Artifact Reduction Using U-Net Model with Three-Dimensional Simulation-Based Datasets for Brain Magnetic Resonance Images
Next Article in Special Issue
Patch-Level Feature Selection for Thoracic Disease Classification by Chest X-ray Images Using Information Bottleneck
Previous Article in Journal
Measurement of the Mapping between Intracranial EEG and fMRI Recordings in the Human Brain
Previous Article in Special Issue
DSP-KD: Dual-Stage Progressive Knowledge Distillation for Skin Disease Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Error Detection on Medical Knowledge Graphs via Intrinsic Label

1
Zhejiang Laboratory, Hangzhou 311121, China
2
School of Information Science and Technology, East China University of Science and Technology, Shanghai 200237, China
*
Author to whom correspondence should be addressed.
Bioengineering 2024, 11(3), 225; https://doi.org/10.3390/bioengineering11030225
Submission received: 26 January 2024 / Revised: 21 February 2024 / Accepted: 24 February 2024 / Published: 27 February 2024
(This article belongs to the Special Issue Artificial Intelligence for Better Healthcare and Precision Medicine)

Abstract

:
The construction of medical knowledge graphs (MKGs) is steadily progressing from manual to automatic methods, which inevitably introduce noise, which could impair the performance of downstream healthcare applications. Existing error detection approaches depend on the topological structure and external labels of entities in MKGs to improve their quality. Nevertheless, due to the cost of manual annotation and imperfect automatic algorithms, precise entity labels in MKGs cannot be readily obtained. To address these issues, we propose an approach named Enhancing error detection on Medical knowledge graphs via intrinsic labEL (EMKGEL). Considering the absence of hyper-view KG, we establish a hyper-view KG and a triplet-level KG for implicit label information and neighborhood information, respectively. Inspired by the success of graph attention networks (GATs), we introduce the hyper-view GAT to incorporate label messages and neighborhood information into representation learning. We leverage a confidence score that combines local and global trustworthiness to estimate the triplets. To validate the effectiveness of our approach, we conducted experiments on three publicly available MKGs, namely PharmKG-8k, DiseaseKG, and DiaKG. Compared with the baseline models, the Precision@K value improved by 0.7%, 6.1%, and 3.6%, respectively, on these datasets. Furthermore, our method empirically showed that it significantly outperformed the baseline on a general knowledge graph, Nell-995.

1. Introduction

A knowledge graph (KG) consists of semantic edges and diverse entities, represented as (h,r,t). KGs can be general, such as YAGO [1] and DBpedia [2], or domain-specific, such as biomedical KG [3] and financial KG [4]. Medical KGs, in particular, exhibit unique structures, contain abundant semantic information [5], and play a crucial role in various healthcare applications, including disease diagnosis [6], drug analysis [7], clinical education [8], and data quality management [9]. However, constructing a medical KG from scratch can be a time-consuming and labor-intensive procedure [10]. Recently, there have been studies that propose constructing medical KGs by either fusing existing KGs [11] or extracting entities related to ICD-9 from Freebase [12]. These approaches aim to streamline the process of constructing medical KGs by leveraging existing resources and extracting relevant entities. However, these methods inevitably introduce noisy triplets due to human errors and imperfect algorithms. For example, the correct triplet (Bill Gates, found, Microsoft) might be mistakenly identified as (Bill Gates, found, Google). This can lead to substantial errors in downstream tasks and applications that rely on the accuracy of the knowledge graph [13].
Traditional knowledge graph representation learning algorithms, such as TransE [14], DistMult [15], and RotatE [16], assume the correctness of all triplets in the KG. However, the absence of error detection mechanisms exposes downstream tasks to significant risks. Therefore, it is crucial to develop an error detection algorithm to ensure the reliability of a medical KG.
As the errors present in KGs can be diverse and their nature may be unknown [17], it is nontrivial to detect noisy triplets in KGs [18]. Recently, some studies have enabled error-aware learning against the noisy triplets [19,20]. In particular, CKRL [19] estimates the triplet with confidence through path and KG embedding. To classify the triplets directly with trustworthiness, KGTtm [21] proposes a model that integrates entity, relationship, and KG global-level information. The state-of-the-art error detection method, CAGED [22], effectively combines KG embedding with contrastive learning. This approach estimates the confidence of triplets both locally and globally, enabling the accurate detection of errors in the KG. By utilizing the additional entities’ attribute information, AKAE [23] enhances the error detection process and improves the accuracy of identifying and addressing nontrivial errors in the KG.
While AKAE has shown promising results, obtaining external attribute information for entities can be challenging. Therefore, we propose an approach that leverages the topological structure of the graph and the rich potential entity information within MKGs to extract intrinsic label information. This approach is more versatile and can be applied to various medical KGs without relying on explicit attributes. To this end, we propose the framework named Enhancing error detection on Medical knowledge graphs via intrinsic labEL (EMKGEL). Our contributions are as follows:
  • Noting the abundance of entities’ labels in medical KGs, we propose a novel method that extracts the intrinsic label information of entities via a hyper-view KG. Further, we establish the hyper-view KG ourselves due to its absence;
  • Aiming to integrate the topological information and intrinsic label information, we propose a hyper-view GAT consisting of a bi-LSTM layer for capturing local structural messages and a modified graph attention mechanism for modeling neighborhood information with potential labels’ messages;
  • Ranking the triplet by confidence score, we conduct comprehensive experiments on three medical KGs and a general KG and outperform other methods.
This research aims to make a valuable contribution to the advancement of medical knowledge graphs within the open-source community. The proposed approach can be effectively utilized for identifying errors in knowledge graph construction. By leveraging this method, it is possible to reduce the time required for construction while simultaneously improving the overall quality of the knowledge graph. In summary, the proposed method represents a promising tool for error detection in knowledge graphs and has wide-ranging implications for advancing the field.

2. Related Work

The related work of the error-detection research community can be divided into three categories: knowledge graph representation learning, error-aware knowledge graph embedding, and knowledge graph error detection.

2.1. Knowledge Graph Representation Learning

Numerous studies have been dedicated to the field of knowledge graph (KG) representation learning, which has emerged as a crucial foundation for various downstream tasks [24,25]. Based on the scoring function used and whether a global graph structure is utilized, KG representation learning methods can be broadly categorized into three groups: translation-based, semantic matching, and GNN-based models.
Translation-based models, exemplified by TransE [14], have garnered widespread attention due to their pioneering work in transforming the KG into a continuous vector space. However, TransE is not well-suited for handling 1-N, N-1, and N-N relations within the knowledge graph. To tackle the issue, several variants of TransE are proposed, including TransH [26], TransR [27], and TransD [28]. In particular, to fit the various relation patterns, including symmetry/antisymmetry, inversion, and composition, RotatE [16] projects the entities and relations into complex vector space via rotation.
Semantic matching models employ a similarity-based scoring function to evaluate the likelihood of the facts. One representative model in this category is RESCAL [29], which utilizes a matrix of full rank to depict relations. Building upon this, DistMult [15] transforms the relation matrix into a diagonal matrix. CompIEx [30] expands on DisMult by incorporating complex space and effectively addressing both symmetry and asymmetry concerns.
GNN-based models integrate neighborhood information to improve the performance of the KG embedding, e.g., ConvKB [31] and CompGCN [32]. Specifically, ConvE [33] utilizes 2D convolution to represent entities and relations. R-GCN [34] obtains the structural information of multi-relation graphs through relation matrices and node neighbors, but the parameters explode as the relations grow. These methods assign the same importance to nodes’ neighbors, while the different nodes make different contributions. To tackle the issue, KGAT [35] introduces an attention mechanism to learn the weight of nodes in a neighborhood.
However, these methods ignore the noisy triplets in KGs, which may decrease the quality of KG representation learning. Thus, it is urgent to develop an algorithm for error-aware KG embedding or to detect errors in the KG.

2.2. Error-Aware Knowledge Graph Embedding

Error-aware KG embedding is end-to-end representation learning that learns the embedding of entities and relations with noise. Aiming to detect possible errors, Dong [36] constructed a knowledge vault by fusing existing KG embeddings with probability scores. Recently, CKRL [19] offers a novel approach that incorporates the confidence score while learning knowledge representations in the KG with noise. While CKRL has shown effectiveness in noise detection by confidence scores, its utilization of uniform negative sampling methods and a strict triple quality function can potentially lead to zero loss problems and false detection. To address these challenges, NKRL [20] introduces the concept of negative confidence and proposes a negative sampling method for training. In addition, TransT [37] measures the confidence score via external entity type and description. These models generate confidence scores in representation learning to estimate the triplets, thereby enhancing the robustness of KG embedding with noise.

2.3. Knowledge Graph Error Detection

Error detection has been a long-standing challenge for many years. Primarily, studies focused on rule-based methods [38], cluster-based methods [39], and distance-based methods [40]. Among these methods, there have been several innovative works for error detection. Undoubtedly, it is straightforward to rank triplets by their embedding score function, e.g., TransE’s score function is h + r t , where the larger the score, the more reliable the triplet. In addition, a classification-based method, KGClean [41], utilizes a novel method called AL-detect to identify errors. Recently, Jia [21] proposed a triple trustworthiness measurement model that combines intra-triplet, structural, and global inference information. This approach enables the fusion of entity, relationship, and path-level information. The state-of-the-art method, CAGED [22], introduces the concept of a link pattern for data augmentation and subsequently integrates KG embedding and contrastive learning to combine local and global information. Based on this, AKAE [23] incorporates the external entity attribute information.

3. Problem Statement

Given a knowledge graph G , we define G = ( E , R , S ) , where E , R , S represent the sets of entity, relations, and triplets, respectively. The major notations used in this paper can be followed in Table 1
Definition 1.
Errors in KGs. Considering a triplet (h, r, t), there is an error if there is a mismatch of the head, relation, or tail, represented as (h’, r’, t’). Two examples illustrate this. Firstly, the triplet (Nanjing, is the capital of, China) incorrectly associates Nanjing as the capital of China, whereas the correct head entity should be Beijing. Secondly, in the triplet (Jimmy Carter, is the son of, Chip Carter), the relationship between the two individuals, Jimmy Carter and Chip Carter, is inverted. In real-world knowledge graphs, there are no ground-truth datasets that include errors. To simulate noisy triplets, we introduce errors by randomly replacing either the head or tail of the original triplets. As a result, we obtain modified triplets with noise denoted as S = ( h , r , t ) . It is important to note that entities and relations outside of the knowledge graph are not considered in this process.
Definition 2.
The triplet-level KG G ^ . We consider a triplet ( h , r , t ) as a node, and we define a relationship between two triplets that share either the head entity or the tail entity, e.g., there is a link between ( e i , r i j , e j ) and ( e j , r j k , e k ) . We denote G ^ = ( S , V , T ) , where S represents the set of triplets from the original knowledge graph G , V represents the edges connecting the triplets that share entities, and T represents the set of new triplets introduced in the process. As illustrated in Figure 1b, X 1 and X 3 share the entity Alzheimer’s disease, so they have a hidden connection.
Definition 3.
The hyper-view KG G . For any given triplet ( h , r , t ) in G that shares the same ( h , r ) pair, we define H t = { t i | ( h , r , t i ) S } as the set of all tail entities associated with this pair. Similarly, we define H h = { h i | ( h i , r , t ) S } as the set of all head entities that share the same relation r and tail t. As depicted in Figure 1a, the head entities at the top share the same couple (belong department, Neurology). Entities such as cerebral infarction, NeuroLyme disease, syringomyelia, and Alzheimer’s disease are likely to be labeled as diseases. Similarly, chestnut and ginkgo are likely to be labeled as food items.
Definition 4.
Confidence score [19]. To classify a triplet, we introduce an estimated score ranging from 0 to 1. For a true triplet (h, r, t), the confidence score is expected to be close to 1, indicating a high degree of confidence in its accuracy.
Unless otherwise specified, the boldface notation in the formulas represents vectors.

4. Methodology

Traditional methods for KG representation learning typically approach the problem by modeling the KG as a heterogeneous graph and learning embeddings for entities (nodes) and relations (edges) [14,26,27]. However, most algorithms struggle to handle the complex relations between triplets. To tackle this issue, several studies [32,33,34,35] propose methods that aggregate information from neighborhood nodes for latent messages. Furthermore, some studies [23,37] explore the integration of external information, such as entities’ labels, to enhance KG embedding. However, obtaining label information for entities in MKGs is not always a straightforward process. Thus, this paper proposes a novel framework named Enhancing error detection on Medical knowledge graphs via intrinsic labEL (EMKGEL). Inspired by recent work [22,23], we construct a triplet-level KG in Definition 2 to capture neighborhood information. In contrast to CAGED, which employs a hyper-parameter for error filtering, our approach involves extracting potential label information at a hyper-view level. We then integrate this information into the attention mechanisms within our framework. Our intuition is that a triplet can acquire analogous label feature information from the set associated with the same hyperedge. For example, the subclass words under the same parent-class words often share similar semantic contexts.
As illustrated in Figure 2, our proposed EMKGEL consists of a multi-view KG, a hyper-view GAT, and a joint confidence score. In our proposed approach, we first generate a triplet-level KG and a hyper-view KG, as described in Definitions 2 and 3. We then employ an attention mechanism to capture the intrinsic label information of entities in the hyper-view KG. This information is subsequently incorporated into the triplets using the hyper-view GAT to enhance the nodes. To train the model, we utilize a combination of KG embedding loss and global triplet embedding loss. This joint training enables the model to learn meaningful representations that capture both the structural relationships within the KG and the semantic contexts of the triplets. Finally, we estimate triplets using a joint confidence score based on the learned representations.

4.1. Hyper-View GAT for Representation Learning

Noisy triplets in KGs can have a detrimental effect on representation learning, thereby jeopardizing downstream tasks. It is crucial to ensure the reliability of the encoder to mitigate the impact of these errors. To address this challenge, CAGED introduces an error-aware GNN that filters out errors. However, determining an optimal hyper-parameter for error filtering is a complex task, making it challenging to apply CAGED to different KGs effectively. In this paper, we present a novel method to extract intrinsic label information from the hyper-view KG. We leverage this information to enhance the representation of triplets by incorporating it into a hyper-view GAT. By combining local structural information and neighborhood triplet messages, the hyper-view GAT effectively integrates multiple sources of information. Following the integration process, we estimate the reliability of each triplet by assigning a confidence score. This score serves as a measure of the triplet’s quality, aiding in the interpretation and utilization of the KG data.
Figure 2. For training purposes, all triplets are generated with pairwise negative examples, denoted by red. This involves randomly replacing the head entity, the relation, and the tail entity in each triplet. (a) We construct a triplet-level KG and a hyper-view KG. (b) Enhanced by intrinsic label information, triplet-level nodes learn the embedding from neighborhood messages. (c) After training on a joint loss, we estimate the confidence of the combined local and global trustworthiness of triplets.
Figure 2. For training purposes, all triplets are generated with pairwise negative examples, denoted by red. This involves randomly replacing the head entity, the relation, and the tail entity in each triplet. (a) We construct a triplet-level KG and a hyper-view KG. (b) Enhanced by intrinsic label information, triplet-level nodes learn the embedding from neighborhood messages. (c) After training on a joint loss, we estimate the confidence of the combined local and global trustworthiness of triplets.
Bioengineering 11 00225 g002

4.1.1. Local Structural Information Modeling

Applying a GAT [42] for learning emphasizes the capture of latent information from neighborhood nodes. However, this approach can potentially weaken the inherent information contained within the triplets, denoted as h r t . Taking inspiration from CAGED, we utilize a bidirectional LSTM to acquire local representations that preserve the specific structural information of the triplets within global information learning.
x = [ x h ; x r ; x t ] = b i l s t m ( h , r , t ) .
As shown in Equation (1), we initialize a triplet ( h , r , t ) and pass the resulting vector ( h , r , t ) through the bi-LSTM layer. The output x represents the local triplet embedding, which captures the structural information of the triplet. Subsequently, we utilize this local embedding x as the input for the global modeling layer.

4.1.2. Intrinsic Label Information in Hyper-View KG

In the MKG, there is a wealth of untapped and exploitable unknown entity label information available. To uncover the latent label information of entities without external information, we construct a hyper-view KG that enriches each entity with diverse attributes within the medical domain. For an embedding triplet x , the head entity h and the tail entity t have corresponding sets, as defined in Definition 3. Taking the shared pair ( h , r ) as an example, we can obtain a set H t for the tail entity. Based on our assumption, the entities in H t likely share similar characteristics or properties, allowing us to infer potential labels or attributes for the tail entity based on this set.
To capture the messages associated with potential labels, we employ an attention mechanism,
q t , i = f a t t ( t , t i ) ,
q t , i indicates the coefficients across the original t and the i-th tail entity in H t . f a t t ( · ) denotes a single layer feed-forward neural network.
To reduce bias and ensure a fair comparison among the coefficients, we normalize them using a Softmax function. By incorporating information from the entire set H t , we can capture the potential labels or attributes associated with H t ,
α t , i = e x p ( q t , i ) j = 1 H t e x p ( q t , j ) ,
t * = i = 1 | H t | α t , i · t i .
Similarly, we obtain the h * from the all head entities of H h ,
h * = i = 1 | H h | α h , i · h i .
To estimate the importance of the h i * and t i * , we employ the similarity calculation. The hyper-view score h y p e r x is utilized to gauge the significance of neighboring nodes,
h y p e r x = s i m ( ( h , t ) , ( h * , t * ) ) ,
By considering the hyper-view score, we can determine the level of contribution that the neighboring nodes have towards the representation and understanding of the target triplet.

4.1.3. Neighborhood Information Modeling

Relying solely on KG embeddings to estimate the confidence of a triplet is insufficient. It is crucial to incorporate contextual information from neighboring triplets to enhance the confidence estimation process. While previous methods such as R-GCN [34] and KGAT [35] have shown effectiveness in leveraging neighborhood information, they may encounter a decline in performance when neighboring triplets contain noisy or erroneous messages, as highlighted in the work of CAGED [22]. To address this issue, CAGED introduces a hyper-parameter that helps mitigate the impact of errors. Given the abundant availability of entity label information in MKGs, we utilize hyper-view scores to enhance the neighborhood nodes.
Specifically, for a given anchor triplet x with m neighboring triplets x 1 , x j , , x m , we aggregate the information from these neighbors to update the representation of the anchor triplet. The weights of anchor triplet x and neighboring triplets x j are calculated as follows:
e j = a ( W c x , W c x j ) ,
where e j indicates the weight of x j to x . a indicates the attention function a : R d × R d R , W c R d × R d R is a trainable parameter matrix that projects triplet x into the same vector space.
Then, we incorporate the hyper-view score to enhance the label information, as depicted in Equations (2)–(6). In detail, we make a dot-product between e j and h y p e r x j in normalization,
α j = e x p ( h y p e r x j · e j ) k = 1 m e x p ( h y p e r x k · e k ) ,
where α j indicates the normalized coefficient weight of the j-th triplet x j to anchor triplet x .
Finally, we obtain the reconstructed vector z for the original anchor triplet x . And the σ ( · ) denotes the Sigmoid function,
z = σ ( j = 1 m α j W c x j ) .

4.2. Joint Training Strategy

To capture the semantic and latent information, we introduce a training strategy to integrate KG embedding loss and global embedding loss.
Based on the translation assumption, we utilize the TransE score function d l o c a l to fit the local structural information,
d l o c a l = h + r t .
For neighborhood information embedding, we use the d g l o b a l to estimate the distance between the anchor triplet x and reconstruct triplet z as follows:
d g l o b a l = x z .
To integrate the KG and global triplet embedding, we introduce the trade-off hyper-parameter λ . In Section 5, we specifically investigate the effect of different values of λ . The calculation is as follows:
d j o i n t = d l o c a l + λ · d g l o b a l .
Subsequently, we leverage a margin-based ranking loss function for negative sampling during the training process following the previous work [14],
L = ( h , r , t ) S ( h , r , t ) S [ γ + d j o i n t ( h , r , t ) d j o i n t ( h , r , t ) ] + ,
where [ ] + equals m a x ( 0 , x ) , λ is a hyper-parameter. S indicates the set of the original triplets, and S indicates the set of the negative triplets that randomly replace the head and tail entities. It is crucial to ensure that corrupted triplets are not in S and are non-repetitive.

4.3. Confidence Score

After training, we obtain the confidence score, as depicted in Equation (14):
C ( h , r , t ) = σ ( λ · s i m ( x , z ) h + r t 2 2 ) ,
The σ ( · ) denotes the Sigmoid function. The function s i m ( · ) denotes the similarity of the original triplet x and the reconstructed triplet z . The confidence score ranges from 0 to 1, with higher values indicating a stronger positive correlation for the triplet.
The learning process of our method is summarized in Algorithm 1.
Algorithm 1 Error detection on medical knowledge graphs via intrinsic label information
Input: Knowledge graph G with noise
Output: KG embeddings and confidence score
  1:
Initialize network parameters,
  2:
Construct a triplet-level and a hyper-view KG as per Definitions 2 and 3, respectively.
  3:
while not converged do
  4:
   for each ( h , r , t ) ∈ S do
  5:
   Modeling the local structural information of triplets as defined in Equation (1),
  6:
   Extract the intrinsic label information in hyper-view and then compute the importance score of triplets as defined in Equation (6),
  7:
   Acquire the representation in hyper-view GAT in Equation (9),
  8:
   Compute the KG embedding distance in Equation (10) and global triplet embedding distance in Equation (11). Combined with a trade-off parameter λ and obtain the joint loss in Equation (13).
  9:
end while
10:
Compute the confidence score as defined in Equation (14).

5. Experiments and Discussion

In this section, we will provide detailed experimental settings. Through the parameter analysis, ablation study, and case study, we validate the effectiveness of the proposed method, EMKGEL.

5.1. Experimental Settings

In this section, we provide a detailed overview of the experimental settings, including datasets, baseline methods, and evaluation metrics.

5.1.1. Benchmark Datasets

Similar to prior studies, such as [19,22,23], we adopt the approach of randomly replacing head and tail entities to generate noisy triplets. As depicted in Definition 1, we introduce 5% noisy triplets into three medical real-world KGs and 5%, 10%, and 15% noisy triplets into one general KG to explore the robustness of our method.
PharmKG-8k [11] is a multi-relational attribute biomedical knowledge graph composed of more than 500,000 individual interconnections between genes, drugs, and diseases, with 29 relation types over a vocabulary of 8000 disambiguated entities.
DiaKG is derived from 41 publicly published diabetes guidelines and consensus documents, covering the most extensive range of research topics and hot areas in recent years.
DiseaseKG is a knowledge graph built upon common disease information utilizing the cnSchema framework.
Detailed information on the datasets is summarized in Table 2.

5.1.2. Baseline Methods

In our experiments, we introduce KG embedding baseline methods and error detection baseline methods.
KG embedding: We compare them with the traditional representation learning methods, including TransE [14], DistMult [15], and RotatE [16]. We leverage the function score as the confidence score after training. In TransE, we employ the Euclidean distance | h + r t | 2 as the confidence score.
Error detection: For KG error detection methods, we evaluate the proposed method against state-of-the-art approaches, including CKRL [19], KGTtm [21], and CAGED [22].

5.1.3. Evaluation Metrics

Consistent with previous studies [22,23], we adopt the practice of ranking all triplets based on their confidence score. Triplets with lower scores are considered potential candidates for being noisy triplets. Precision@K and Recall@K metrics are utilized to estimate the effectiveness. In detail, Precision@K denotes the TopK lowest confidence score triplets among the TopK triplets. Recall@K denotes the TopK lowest confidence score triplets among all triplets.
P r e c i s i o n @ K = | Errors Discovered in TopK Ranking list | K
R e c a l l @ K = | Errors Discovered in TopK Ranking list | | Total Number of Errors in KG |

5.1.4. Implementation Details

We conduct experiments on GPU NVIDIA GeForce RTX 3090; the Python version is 3.8 and the Pytorch version is 1.11.0. Based on the average in-degree of the datasets, the number of neighbors for each triplet in the four datasets is set to 59/2/7/2. The embedding hidden size is set to 100, the same as the bi-LSTM hidden size. Default Xavier initialization and an initial learning rate of 0.003 were used.
In our experiments, we explore different hyper-parameters to assess their impact on the results. The trade-off parameter λ is set from 0.001 to 1000, while the margin parameter γ is adjusted within the range of 0 to 1. To mitigate the impact of randomness introduced by erroneous triplets, we average the experimental data across 10 random seeds ranging from 0 to 9.

5.2. Results and Analysis

In this section, we conduct a comprehensive evaluation to assess the effectiveness of our method across four datasets. Through thorough observation and analysis, we demonstrate that our proposed method performs effectively in all four datasets. For clarity, we highlight the optimal results in black and underline the second-best results. Additionally, we indicate that Precision@K is equal to Recall@K when K equals ratios by using an asterisk (*).

5.2.1. Main Results

As depicted in Table 3, the results demonstrate that (1) error detection methods outperform traditional embedding methods on the three medical KGs, and (2) notably, our proposed method, EMKGEL, outperforms all existing methods, delivering the best performance.
Specifically, traditional embedding methods such as TransE and RotatE solely focus on local structural information while neglecting global triplet embedding. Consequently, this limited perspective can result in losing important messages from neighborhood triplets. To tackle this issue, CAGED addresses the limitation by incorporating global triplet embedding through contrastive learning. This integration enables CAGED to capture and leverage the global context of triplets, detecting a higher number of errors in the KG. Different from CAGED’s utilization of the uncertainty parameter μ for error filtering, our approach takes a novel approach. Initially, we leveraged a hyper-view KG to extract potential label information for entities. Subsequently, we estimated the importance of nodes by assigning hyper-view scores, thereby enhancing their impact on neighboring nodes. In our experiments, we observe that our method demonstrates improvements of 0.7%, 6.1%, and 3.6% on PharmKG-8k, DiseaseKG, and DiaKG, respectively.
Additionally, we introduce different ratios of 5%, 10%, and 15% on Nell-995 to observe their effectiveness. As depicted in Table 4, our method consistently achieves the best results across different cases, highlighting the robustness of EMKGEL.

5.2.2. Ablation Study

To validate the individual components of our proposed method, we conduct comprehensive experiments. Firstly, we replace the bi-LSTM by simply concatenating the triplet embedding to assess the impact of local structural information. As shown in Table 5, we can observe that the variant employing only concatenation exhibits inferior performance due to the absence of structural information. Secondly, we replace the hyper-view GAT with a simple GAT. As indicated in Table 5, our hyper-view GAT outperforms the GAT, thereby demonstrating the effectiveness of label information in enhancing the embeddings. Thirdly, we proceed to eliminate the KG and triplet embedding losses individually. Upon analyzing the results presented in Table 5, it becomes apparent that the model’s performance suffers when either the KG or triplet embedding loss is removed. This observation suggests a strong interdependence between KG embedding and triplet embedding, indicating that these two components work collaboratively to enhance the model’s performance. Lastly, we introduce a replacement for the TransE score function, as shown in Equation (10), by adopting the RotatE score function. This modification aims to explore the flexibility of different score functions and their impact on the model’s performance.

5.2.3. Parameter Analysis

The λ is the trade-off parameter that balances the KG embedding (e.g., h + r t ) and global triplet embedding (e.g., x z ). To investigate the impact of λ , we set it from 0.001 to 1000. We conducted experiments on all four datasets and the evaluation on Recall@K. Based on the findings depicted in Figure 3a, we observe that (1) PharmKG-8k, DiaKG, and Nell-995 demonstrate their best performance when the value of λ is set to 10; (2) DiseaseKG achieves its optimal result when λ is set to 0.1; (3) at the outset, as λ increases, the model’s performance shows improvement. However, once it reaches the optimal value, a decline in performance is observed. This suggests that while enhancing the impact of global embedding initially boosts performance, there is a point of diminishing returns. Pushing the value of λ beyond this point does not yield the best performance for the model.
The γ is the margin parameter. As shown in Figure 3b, the optimal result is γ = 0.5 , and the trends of the four datasets are essentially identical.

5.3. Case Study

To investigate how the hyper-score enhances error detection, we conducted a case study on the anchor triplet (Diabetes ketoacidosis, Symptom, Polydipsia) of DiseaseKG. As shown in Figure 4, we present the hyper-scores of neighbors. In our assumption, the hyper-view GAT will enhance the triplets with high hyper-scores and reduce the impact of triplets with low hyper-scores. Finally, we assume that the confidence score will significantly differentiate between true triples and noisy triples.
To confirm the results, we set the number of share(h, r)/share(r, t) of X 1 to X 5 , and the confidence score of each triplet, respectively.
As shown in Table 6, we observe that our proposed method outperforms CAGED. In detail, CAGED cannot lower the confidence score on three noisy triplets, while ours does so and to a better degree. Furthermore, X 2 , X 5 , and X 3 possess more intrinsic label information, as indicated by their hyper-scores of {0.8993, 0.9103, 0.5371}, due to a high number of share(h, r) and share(r, t). Hyper-view GAT enhances the representation of X 2 , X 5 , and X 3 . Conversely, X 1 and X 4 contain less intrinsic information as they have a low number of share(h, r) and share(r, t). Consequently, hyper-view GAT diminishes the influence of X 1 and X 4 . However, the confidence score of noisy triplet X 3 (Diabetes ketoacidosis, Medication, Shagliptin tablets) in both methods is ambiguous. In real-world scenarios, it is not enough to prescribe medication based solely on simple triplet information; it is also necessary to consider the actual situation. Therefore, this is a common issue with existing methods based on knowledge graphs.
To show a straightforward validation of our method, we present a visualization in Figure 5. The y-axis represents the confidence scores assigned to the triplets. True triplets are denoted by green nodes, while false triplets are represented by red nodes. In comparison to CAGED in Figure 5b, our model in Figure 5a assigns lower confidence scores to noisy triplets, approaching zero. This visualization serves as evidence of the effectiveness of our model in real-world scenarios.

6. Conclusions

In this paper, we propose a novel framework named Enhancing error detection on Medical knowledge graphs via intrinsic labEL (EMKGEL). Firstly, we construct a hyper-view KG and a triplet-level KG. The former aims to capture intrinsic label information, and the latter focuses on neighboring information. Secondly, we introduce the hyper-view GAT to incorporate the entity label information into the triplet. Then, we integrate KG embedding and global triplet embedding in the training stage. In the end, we estimate each triplet by their confidence score. The evaluation on three medical KGs and one general KG demonstrates the effectiveness of EMKGEL. We believe that our method can be an effective tool for error detection during KG construction.
However, there remain some challenges that future work needs to address:
  • Existing error detection methods [19,21,22,23] only take into account the entities and relations already present in the knowledge graph, while error triplets could originate from outside the dataset [43]. To address the limitation, combining textual information from large language models with graph structure information is a promising direction [44,45].
  • The currently used max dataset size is approximately 500,000, and in the future, we will consider conducting more experiments on large-scale graphs while maintaining the effectiveness and reducing the training time.
  • We will explore more meaningful downstream tasks, such as knowledge-based question answering in the medical field.

Author Contributions

Conceptualization, G.Y., Q.Y. and T.R.; methodology, G.Y.; software, G.Y.; validation, G.Y.; formal analysis, G.Y.; investigation, G.Y.; resources, G.Y., Q.Y. and T.R.; data curation, G.Y.; writing—original draft preparation, G.Y.; writing—review and editing, G.Y., Q.Y. and T.R.; visualization, G.Y.; supervision, G.Y., Q.Y. and T.R.; project administration, G.Y., Q.Y. and T.R.; funding acquisition, Q.Y. and T.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Program of China (2021YFC2701800, 2021YFC2701801) and ZheJiang Lab (No. 2019ND0AB01).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets we used in experiments are available. PharmKG-8K (https://github.com/biomed-AI/PharmKG, accessed on 10 December 2023), DiseaseKG (http://openkg.cn/dataset/disease-information, accessed on 10 December 2023), DiaKG (http://openkg.cn/dataset/diakg, accessed on 10 December 2023), Nell-995 (https://paperswithcode.com/dataset/nell-995, accessed on 10 December 2023).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Suchanek, F.M.; Kasneci, G.; Weikum, G. Yago: A core of semantic knowledge. In Proceedings of the 16th International Conference on World Wide Web, Banff, AB, Canada, 8–12 May 2007; pp. 697–706. [Google Scholar]
  2. Lehmann, J.; Isele, R.; Jakob, M.; Jentzsch, A.; Kontokostas, D.; Mendes, P.N.; Hellmann, S.; Morsey, M.; Van Kleef, P.; Auer, S.; et al. Dbpedia—A large-scale, multilingual knowledge base extracted from wikipedia. Semant. Web 2015, 6, 167–195. [Google Scholar] [CrossRef]
  3. Yuan, J.; Jin, Z.; Guo, H.; Jin, H.; Zhang, X.; Smith, T.; Luo, J. Constructing biomedical domain-specific knowledge graph with minimum supervision. Knowl. Inf. Syst. 2020, 62, 317–336. [Google Scholar] [CrossRef]
  4. Zehra, S.; Mohsin, S.F.M.; Wasi, S.; Jami, S.I.; Siddiqui, M.S.; Syed, M.K.U.-R.R. Muhammad Khaliq-Ur-Rahman Raazi Syed Muhammad Shoaib Siddiqui Financial knowledge graph based financial report query system. IEEE Access 2021, 9, 69766–69782. [Google Scholar] [CrossRef]
  5. Finlayson, S.G.; LePendu, P.; Shah, N.H. Building the graph of medicine from millions of clinical narratives. Sci. Data 2014, 1, 1–9. [Google Scholar] [CrossRef] [PubMed]
  6. Papageorgiou, E.I.; Huszka, C.; De Roo, J.; Douali, N.; Jaulent, M.C.; Colaert, D. Application of probabilistic and fuzzy cognitive approaches in semantic web framework for medical decision support. Comput. Methods Programs Biomed. 2013, 112, 580–598. [Google Scholar] [CrossRef]
  7. Mohamed, S.K.; Nováček, V.; Nounu, A. Discovering protein drug targets using knowledge graph embeddings. Bioinformatics 2020, 36, 603–610. [Google Scholar] [CrossRef]
  8. Ettorre, A.; Rocha Rodríguez, O.; Faron, C.; Michel, F.; Gandon, F. A knowledge graph enhanced learner model to predict outcomes to questions in the medical field. In Proceedings of the International Conference on Knowledge Engineering and Knowledge Management, Bolzano, Italy, 16–20 September 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 237–251. [Google Scholar]
  9. Fei, X.; Chen, P.; Wei, L.; Huang, Y.; Xin, Y.; Li, J. Quality Management of Pulmonary Nodule Radiology Reports Based on Natural Language Processing. Bioengineering 2022, 9, 244. [Google Scholar] [CrossRef]
  10. Blagec, K.; Barbosa-Silva, A.; Ott, S.; Samwald, M. A curated, ontology-based, large-scale knowledge graph of artificial intelligence tasks and benchmarks. Sci. Data 2022, 9, 322. [Google Scholar] [CrossRef] [PubMed]
  11. Zheng, S.; Rao, J.; Song, Y.; Zhang, J.; Xiao, X.; Fang, E.F.; Yang, Y.; Niu, Z. PharmKG: A dedicated knowledge graph benchmark for bomedical data mining. Brief. Bioinform. 2021, 22, bbaa344. [Google Scholar] [CrossRef]
  12. Teng, F.; Yang, W.; Chen, L.; Huang, L.; Xu, Q. Explainable prediction of medical codes with knowledge graphs. Front. Bioeng. Biotechnol. 2020, 8, 867. [Google Scholar] [CrossRef]
  13. Heindorf, S.; Potthast, M.; Stein, B.; Engels, G. Vandalism detection in wikidata. In Proceedings of the 25th ACM International on Conference on Information and Knowledge Management, Indianapolis, IN, USA, 24–28 October 2016; pp. 327–336. [Google Scholar]
  14. Bordes, A.; Usunier, N.; Garcia-Duran, A.; Weston, J.; Yakhnenko, O. Translating embeddings for modeling multi-relational data. Adv. Neural Inf. Process. Syst. 2013, 26, 2787–2795. [Google Scholar]
  15. Yang, B.; Yih, W.T.; He, X.; Gao, J.; Deng, L. Embedding entities and relations for learning and inference in knowledge bases. arXiv 2014, arXiv:1412.6575. [Google Scholar]
  16. Sun, Z.; Deng, Z.H.; Nie, J.Y.; Tang, J. Rotate: Knowledge graph embedding by relational rotation in complex space. arXiv 2019, arXiv:1902.10197. [Google Scholar]
  17. Melo, A.; Paulheim, H. Detection of relation assertion errors in knowledge graphs. In Proceedings of the Knowledge Capture Conference, Austin, TX, USA, 4–6 December 2017; pp. 1–8. [Google Scholar]
  18. Abedini, F.; Keyvanpour, M.R.; Menhaj, M.B. Correction Tower: A general embedding method of the error recognition for the knowledge graph correction. Int. J. Pattern Recognit. Artif. Intell. 2020, 34, 2059034. [Google Scholar] [CrossRef]
  19. Xie, R.; Liu, Z.; Lin, F.; Lin, L. Does william shakespeare really write hamlet? Knowledge representation learning with confidence. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32, p. 11924. [Google Scholar]
  20. Shan, Y.; Bu, C.; Liu, X.; Ji, S.; Li, L. Confidence-aware negative sampling method for noisy knowledge graph embedding. In Proceedings of the 2018 IEEE International Conference on Big Knowledge (ICBK), Singapore, 17–18 November 2018; pp. 33–40. [Google Scholar]
  21. Jia, S.; Xiang, Y.; Chen, X.; Wang, K. Triple trustworthiness measurement for knowledge graph. In Proceedings of the World Wide Web Conference, San Francisco, CA, USA, 13–17 May 2019; pp. 2865–2871. [Google Scholar]
  22. Zhang, Q.; Dong, J.; Duan, K.; Huang, X.; Liu, Y.; Xu, L. Contrastive knowledge graph error detection. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management, Atlanta, GA, USA, 17–21 October 2022; pp. 2590–2599. [Google Scholar]
  23. Zhang, Q.; Dong, J.; Tan, Q.; Huang, X. Integrating entity attributes for error-aware knowledge graph embedding. IEEE Trans. Knowl. Data Eng. 2023. [Google Scholar] [CrossRef]
  24. Li, Q.; Li, L.; Zhong, J.; Huang, L.F. Real-time sepsis severity prediction on knowledge graph deep learning networks for the intensive care unit. J. Vis. Commun. Image Represent. 2020, 72, 102901. [Google Scholar] [CrossRef]
  25. Jiang, J.; Wang, H.; Xie, J.; Guo, X.; Guan, Y.; Yu, Q. Medical knowledge embedding based on recursive neural network for multi-disease diagnosis. Artif. Intell. Med. 2020, 103, 101772. [Google Scholar] [CrossRef]
  26. Wang, Z.; Zhang, J.; Feng, J.; Chen, Z. Knowledge graph embedding by translating on hyperplanes. In Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence, Québec City, QC, Canada, 27–31 July 2014; Volume 28, pp. 1112–1119. [Google Scholar]
  27. Lin, Y.; Liu, Z.; Sun, M.; Liu, Y.; Zhu, X. Learning entity and relation embeddings for knowledge graph completion. In Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015; Volume 29, pp. 2181–2187. [Google Scholar]
  28. Ji, G.; He, S.; Xu, L.; Liu, K.; Zhao, J. Knowledge graph embedding via dynamic mapping matrix. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Beijing, China, 26–31 July 2015; pp. 687–696. [Google Scholar]
  29. Nickel, M.; Tresp, V.; Kriegel, H.P. A three-way model for collective learning on multi-relational data. In Proceedings of the 28th International Conference on Machine Learning, Bellevue, WA, USA, 28 June–2 July 2011; Volume 11, pp. 809–816. [Google Scholar]
  30. Trouillon, T.; Welbl, J.; Riedel, S.; Gaussier, É.; Bouchard, G. Complex embeddings for simple link prediction. In Proceedings of the 33rd International Conference on Machine Learning, PMLR, New York, NY, USA, 19–24 June 2016; pp. 2071–2080. [Google Scholar]
  31. Nguyen, D.Q.; Nguyen, T.D.; Nguyen, D.Q.; Phung, D. A novel embedding model for knowledge base completion based on convolutional neural network. arXiv 2017, arXiv:1712.02121. [Google Scholar]
  32. Vashishth, S.; Sanyal, S.; Nitin, V.; Talukdar, P. Composition-based multi-relational graph convolutional networks. arXiv 2019, arXiv:1911.03082. [Google Scholar]
  33. Dettmers, T.; Minervini, P.; Stenetorp, P.; Riedel, S. Convolutional 2d knowledge graph embeddings. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32, pp. 1811–1818. [Google Scholar]
  34. Schlichtkrull, M.; Kipf, T.N.; Bloem, P.; Van Den Berg, R.; Titov, I.; Welling, M. Modeling relational data with graph convolutional networks. In Proceedings of the The Semantic Web: 15th International Conference, ESWC 2018, Heraklion, Greece, 3–7 June 2018; Proceedings 15. Springer: Berlin/Heidelberg, Germany, 2018; pp. 593–607. [Google Scholar]
  35. Nathani, D.; Chauhan, J.; Sharma, C.; Kaul, M. Learning attention-based embeddings for relation prediction in knowledge graphs. arXiv 2019, arXiv:1906.01195. [Google Scholar]
  36. Dong, X.; Gabrilovich, E.; Heitz, G.; Horn, W.; Lao, N.; Murphy, K.; Strohmann, T.; Sun, S.; Zhang, W. Knowledge vault: A web-scale approach to probabilistic knowledge fusion. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 601–610. [Google Scholar]
  37. Zhao, Y.; Feng, H.; Gallinari, P. Embedding learning with triple trustiness on noisy knowledge graph. Entropy 2019, 21, 1083. [Google Scholar] [CrossRef]
  38. Ma, Y.; Gao, H.; Wu, T.; Qi, G. Learning disjointness axioms with association rule mining and its application to inconsistency detection of linked data. In Proceedings of the Semantic Web and Web Science: 8th Chinese Conference, CSWS 2014, Wuhan, China, 8–12 August 2014; Revised Selected Papers 8. Springer: Berlin/Heidelberg, Germany, 2014; pp. 29–41. [Google Scholar]
  39. Wang, X.; Wang, X.L.; Wilkes, D.M. A minimum spanning tree-inspired clustering-based outlier detection technique. In Proceedings of the Advances in Data Mining. Applications and Theoretical Aspects: 12th Industrial Conference, ICDM 2012, Berlin, Germany, 13–20 July 2012; Proceedings 12. Springer: Berlin/Heidelberg, Germany, 2012; pp. 209–223. [Google Scholar]
  40. Debattista, J.; Lange, C.; Auer, S. A preliminary investigation towards improving linked data quality using distance-based outlier detection. In Proceedings of the Semantic Technology: 6th Joint International Conference, JIST 2016, Singapore, 2–4 November 2016; Revised Selected Papers 6. Springer: Berlin/Heidelberg, Germany, 2016; pp. 116–124. [Google Scholar]
  41. Ge, C.; Gao, Y.; Weng, H.; Zhang, C.; Miao, X.; Zheng, B. Kgclean: An embedding powered knowledge graph cleaning framework. arXiv 2020, arXiv:2004.14478. [Google Scholar]
  42. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
  43. Xue, B.; Zou, L. Knowledge graph quality management: A comprehensive survey. IEEE Trans. Knowl. Data Eng. 2022, 35, 4969–4988. [Google Scholar] [CrossRef]
  44. Yao, L.; Mao, C.; Luo, Y. KG-BERT: BERT for knowledge graph completion. arXiv 2019, arXiv:1909.03193. [Google Scholar]
  45. Zhang, Y.; Chen, Z.; Zhang, W.; Chen, H. Making Large Language Models Perform Better in Knowledge Graph Completion. arXiv 2023, arXiv:2310.06671. [Google Scholar]
Figure 1. The medical KG in hyper-view and triplet-level. (a) The hyper-view KG. * represents the example template: entities are represented by yellow, while relations are denoted by green. Within the hyper-view, potential label information is highlighted in black. Specifically, tail entities sharing the same couple (Alzheimer’s disease psychosis, recommend) have a hidden potential label as food. Head entities sharing the same couple (belong department, neurology) have a potential label as disease. (b) The triplet-level KG. We use X 1 , X 2 , X 3 to represent triplets. Triplets’ shared entities have hidden connections. Aggregating the neighborhood triplets X 2 , X 3 of the triplet X 1 , we reconstruct it as Z 1 .
Figure 1. The medical KG in hyper-view and triplet-level. (a) The hyper-view KG. * represents the example template: entities are represented by yellow, while relations are denoted by green. Within the hyper-view, potential label information is highlighted in black. Specifically, tail entities sharing the same couple (Alzheimer’s disease psychosis, recommend) have a hidden potential label as food. Head entities sharing the same couple (belong department, neurology) have a potential label as disease. (b) The triplet-level KG. We use X 1 , X 2 , X 3 to represent triplets. Triplets’ shared entities have hidden connections. Aggregating the neighborhood triplets X 2 , X 3 of the triplet X 1 , we reconstruct it as Z 1 .
Bioengineering 11 00225 g001
Figure 3. Impact of hyper-parameters on the four datasets.
Figure 3. Impact of hyper-parameters on the four datasets.
Bioengineering 11 00225 g003
Figure 4. Case study on DiseaseKG. The center triplet of subgraph is (Diabetes ketoacidosis, Symptom, Polydipsia). The other triplets are neighbors of the anchor triplets, which share the same head entity or tail entity, as defined in Definition 2. Detailed triplets are provided on the right side.
Figure 4. Case study on DiseaseKG. The center triplet of subgraph is (Diabetes ketoacidosis, Symptom, Polydipsia). The other triplets are neighbors of the anchor triplets, which share the same head entity or tail entity, as defined in Definition 2. Detailed triplets are provided on the right side.
Bioengineering 11 00225 g004
Figure 5. Scatter of confidence score on DiseaseKG.The red nodes denote noisy triplets and green nodes denote true triplets.
Figure 5. Scatter of confidence score on DiseaseKG.The red nodes denote noisy triplets and green nodes denote true triplets.
Bioengineering 11 00225 g005
Table 1. Major notations.
Table 1. Major notations.
NotationDescription
G the original knowledge graph
(h, r, t)a triplet with the head, relation, and tail
G ^ the triplet-level KG constructed from G
G the hyper-view KG constructed from G
H t the set of the tail entities share the same (h, r)
xrepresentation of the triplet
zrepresentation learned from x
C ( h , r , t ) confidence score of the triplet (h, r, t)
Table 2. The statistical information of datasets.
Table 2. The statistical information of datasets.
DatasetsEntitiesRelationsTripletsIn-Degree
PharmKG-8k726228500,95858.86
DiaKG26581540991.55
DiseaseKG43,97212312,1596.74
Nell-99575,492200154,2131.98
Table 3. Error detection results on three datasets with noisy ratio = 5%. Optimal results are highlighted in black, while the second-best results are underlined. † denotes the result is reproduced by ourselves. * denotes the Precision@K is equal to Recall@K.
Table 3. Error detection results on three datasets with noisy ratio = 5%. Optimal results are highlighted in black, while the second-best results are underlined. † denotes the result is reproduced by ourselves. * denotes the Precision@K is equal to Recall@K.
DatasetsPharmKG-8kDiseaseKGDiaKG
KK = 1%K = 2%K = 3%K = 4%K = 5%*K = 1%K = 2%K = 3%K = 4%K = 5% *K = 1%K = 2%K = 3%K = 4%K = 5% *
Precision@KTransE [14] †0.8120.6810.5840.5150.4610.6230.4980.4250.3700.3310.3410.2800.2270.2070.175
DistMult [15] †0.8020.7220.6640.6070.5470.7540.6360.5330.4580.4020.4390.3040.2520.2190.200
RotatE [16] †0.8710.7620.6810.6090.5630.7270.6010.5010.4270.3740.4870.3170.2600.2250.195
CAGED [22] †0.9580.9170.8450.7610.6740.9360.8700.7760.6870.6110.5620.4590.4000.3710.326
EMKGEL(our)0.9630.9230.8520.7630.6790.9620.8990.8130.7270.6480.5960.4870.4180.3780.338
Recall@KTransE [14] †0.1620.2720.3500.4110.4610.1240.1990.2550.2960.3310.0680.1120.1360.1650.175
DistMult [15] †0.1600.2880.3980.4860.5470.1500.2540.3190.3670.4020.0870.1220.1510.1750.200
RotatE [16] †0.1740.3040.4080.4870.5630.1420.2410.3010.3420.3740.0970.1260.1560.1800.195
CAGED [22] †0.1910.3660.5070.6080.6740.1870.3470.4650.5500.6110.1120.1830.2400.2970.326
EMKGEL(our)0.1920.3690.5110.6110.6790.1920.3590.4870.5810.6480.1190.1940.2510.3020.338
Table 4. Error detection results on Nell-995 with different ratios. Optimal results are highlighted in black, while the second-best results are underlined. † denotes the result is reproduced by ourselves.
Table 4. Error detection results on Nell-995 with different ratios. Optimal results are highlighted in black, while the second-best results are underlined. † denotes the result is reproduced by ourselves.
Ratio5%10%15%
KK = 5%K = 10%K = 15%K = 5%K = 10%K = 15%K = 5%K = 10%K = 15%
Precision@K
TransE [14] †0.3580.2550.1960.5460.4320.3560.6660.5690.490
DistMult [15] †0.2860.2350.2050.4890.4210.3750.6050.5400.494
RotatE [16] †0.3520.2370.1930.4540.3780.3240.6260.5440.474
CKRL [22]0.4500.3060.2360.6790.5240.4210.7450.6460.560
KGTtm [22]0.4810.3200.2420.7130.5270.4370.7880.6730.576
CAGED [22]0.5160.3250.2510.7990.5850.4580.8230.7290.599
EMKGEL(our)0.5420.3440.2640.8070.6060.4880.8420.7320.626
Recall@K
TransE [14] †0.3580.5110.5890.2730.4320.5340.2220.3790.490
DistMult [15] †0.2860.4710.6160.2440.4210.5620.2010.3600.494
RotatE [16] †0.3520.4740.5790.2270.3780.4870.2080.3630.474
CKRL [22]0.4500.6120.7080.3400.5240.6320.2480.4310.560
KGTtm [22]0.4810.6400.7260.3570.5270.6560.2630.4490.576
CAGED [22]0.5160.6500.7530.4000.5850.6870.2740.4860.599
EMKGEL(our)0.5420.6890.7920.4030.6060.7320.2810.4880.732
Table 5. Ablation study on Nell-995 with noisy Ratio = 5%. Optimal results are highlighted in black.
Table 5. Ablation study on Nell-995 with noisy Ratio = 5%. Optimal results are highlighted in black.
Precision@KRecall@K
KK = 1%K = 2%K = 3%K = 4%K = 5%K = 1%K = 2%K = 3%K = 4%K = 5%
EMKGEL0.8870.7920.6920.6130.5420.1770.3170.4150.4900.542
EMKGEL w/o bi-LSTM0.7930.7110.6210.5490.4920.1580.2840.3730.4390.492
EMKGEL w/o HyGAT0.7980.7210.6300.5610.5070.1590.2880.3780.4480.507
EMKGEL w/o local0.8660.7590.6580.5740.5070.1730.3030.3940.4590.507
EMKGEL w/o global0.8280.7540.6770.5940.5180.1650.3010.4060.4750.518
EMKGEL w RotatE0.8850.7680.6610.5830.5230.1770.3070.3970.4670.523
Table 6. The information of neighbors of the anchor triplet. The red color denotes bad case.
Table 6. The information of neighbors of the anchor triplet. The red color denotes bad case.
TripletNumber of Share(h, r)Number of Share(r, t)Confidence Score
GroundedOurCAGED
X669True0.87030.8549
X 2 669True0.83010.7924
X 5 6611True0.92010.8773
X 3 6610Noisy0.65210.4891
X 1 52Noisy0.24310.6312
X 4 41Noisy0.10030.5455
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yu, G.; Ye, Q.; Ruan, T. Enhancing Error Detection on Medical Knowledge Graphs via Intrinsic Label. Bioengineering 2024, 11, 225. https://doi.org/10.3390/bioengineering11030225

AMA Style

Yu G, Ye Q, Ruan T. Enhancing Error Detection on Medical Knowledge Graphs via Intrinsic Label. Bioengineering. 2024; 11(3):225. https://doi.org/10.3390/bioengineering11030225

Chicago/Turabian Style

Yu, Guangya, Qi Ye, and Tong Ruan. 2024. "Enhancing Error Detection on Medical Knowledge Graphs via Intrinsic Label" Bioengineering 11, no. 3: 225. https://doi.org/10.3390/bioengineering11030225

APA Style

Yu, G., Ye, Q., & Ruan, T. (2024). Enhancing Error Detection on Medical Knowledge Graphs via Intrinsic Label. Bioengineering, 11(3), 225. https://doi.org/10.3390/bioengineering11030225

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop