Command Palette
Search for a command to run...
Robert Bamler; Farnood Salehi; Stephan Mandt

Abstract
Knowledge graph embeddings rank among the most successful methods for link prediction in knowledge graphs, i.e., the task of completing an incomplete collection of relational facts. A downside of these models is their strong sensitivity to model hyperparameters, in particular regularizers, which have to be extensively tuned to reach good performance [Kadlec et al., 2017]. We propose an efficient method for large scale hyperparameter tuning by interpreting these models in a probabilistic framework. After a model augmentation that introduces per-entity hyperparameters, we use a variational expectation-maximization approach to tune thousands of such hyperparameters with minimal additional cost. Our approach is agnostic to details of the model and results in a new state of the art in link prediction on standard benchmark data.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| link-prediction-on-fb15k | DistMult (after variational EM) | Hits@10: 0.914 MRR: 0.841 |
| link-prediction-on-fb15k-237 | DistMult (after variational EM) | Hits@10: 0.548 MRR: 0.357 |
| link-prediction-on-wn18 | DistMult (after variational EM) | MRR: 0.911 |
| link-prediction-on-wn18rr | DistMult (after variational EM) | MRR: 0.455 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.