Command Palette
Search for a command to run...
Orthogonal Relation Transforms with Graph Context Modeling for Knowledge Graph Embedding
Yun Tang Jing Huang Guangtao Wang Xiaodong He Bowen Zhou

Abstract
Translational distance-based knowledge graph embedding has shown progressive improvements on the link prediction task, from TransE to the latest state-of-the-art RotatE. However, N-1, 1-N and N-N predictions still remain challenging. In this work, we propose a novel translational distance-based approach for knowledge graph link prediction. The proposed method includes two-folds, first we extend the RotatE from 2D complex domain to high dimension space with orthogonal transforms to model relations for better modeling capacity. Second, the graph context is explicitly modeled via two directed context representations. These context representations are used as part of the distance scoring function to measure the plausibility of the triples during training and inference. The proposed approach effectively improves prediction accuracy on the difficult N-1, 1-N and N-N cases for knowledge graph link prediction task. The experimental results show that it achieves better performance on two benchmark data sets compared to the baseline RotatE, especially on data set (FB15k-237) with many high in-degree connection nodes.
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| link-prediction-on-fb15k-237 | GC-OTE | Hits@1: 0.267 Hits@10: 0.550 Hits@3: 0.396 MR: 154 MRR: 0.361 |
| link-prediction-on-wn18rr | GC-OTE | Hits@1: 0.442 Hits@10: 0.583 Hits@3: 0.511 MR: 2715 MRR: 0.491 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.