Discriminator-based adversarial networks for knowledge graph completion

Document Type

Article

Source of Publication

Neural Computing and Applications

Publication Date

1-1-2022

Abstract

Knowledge graphs (KGs) inherently lack reasoning ability which limits their effectiveness for tasks such as question–answering and query expansion. KG embedding (KGE) is a predominant approach where proximity between relations and entities in the embedding space is used for reasoning over KGs. Most existing KGE approaches use structural information of triplets and disregard contextual information which could be crucial to learning long-term relations between entities. Moreover, KGE approaches mostly use discriminative models which require both positive and negative samples to learn a decision boundary. KGs, by contrast, contain only positive samples, necessitating that negative samples are generated by replacing the head/tail of predicates with randomly chosen entities. They are thus usually irrational and easily discriminable from positive samples, which can prevent the learning of sufficiently robust classifiers. To address the shortcomings, we propose to learn contextualized KGE using pre-trained adversarial networks. We assume multi-hop relational paths(mh-RPs) as textual sequences for competitively learning discriminator-based KGE against the negative mh-RP generator. We use a pre-trained ELECTRA model and feed it with relational paths. We employ a generator to corrupt randomly chosen entities with plausible alternatives and a discriminator to predict whether an entity is corrupted or not. We perform experiments on multiple benchmark knowledge graphs, and the results show that our proposed KG-ELECTRA model outperforms BERT in link prediction.

ISSN

0941-0643

Publisher

Springer Science and Business Media LLC

Disciplines

Computer Sciences

Keywords

Knowledge graph completion, Pre-trained language model, Transformer model

Scopus ID

85135774265

Indexed in Scopus

yes

Open Access

no

Share

COinS