NLP and others
02 May 2021
Keywords: NLP, Link Prediction, Embedding
Link Prediction is a task of Knowledge Graph (KG), and its definition is as follows. KG is composed of many triples, where each triple is represented by (head, relation, tail), and the goal of Link Prediction is to predict new triples using given triples to make a complete KG. This paper proposed an embedding model SimplE based on the Canonical Polyadic (CP) decomposition method which is one of tensor factorization methods. According to the authors, while SimplE overperforms many existing SOTA models based on tensor factorization approach, it is ‘simpler’, interpretable and expressive.
Canonical Polyadic (CP) decomposition is among the first tensor factorization approaches. CP generally performs poorly for link prediction as it learns two independent embedding vectors for each entity, whereas they are really tied.
Proposition 1. For any ground truth over entities $E$ and relations $R$ containing $γ$ true facts, there exists a SimplE model with embedding vectors of size $min(|E| · |R|, γ + 1)$ that represents that ground truth.
The paper gives other examples of the expressiveness of other existing models that might help for understanding: “DistMult is not fully expressive as it forces relations to be symmetric. It has been shown in that ComplEx is fully expressive with embeddings of length at most $|E|{\cdot}|R|$.”
In SimplE, each element of the embedding vector is considered as a feature, and each one represents how important that feature is to the relation.
The implemented properties are:
The advantages of incorporating operations are:
The reason why the paper could achieve fast computation speed by two to four times over the contrasting model ComplEx is that, it removed redundant elements in various ways, including the techniques mentioned above.
The experimental results shows the performance of SimplE is comparable to or superior to the existing models, while faster and more efficient learning is possible.
Nowadays, language models such as GPT-2 and GPT-3, which are trained with billion-parameter or even trillion-parameter that requires enormous computation resources, are in the spotlight in NLP research.
However, that also led many AI research to focus on conducting lightweight and efficient models capable of fast learning with a small amount of resources.