Publication
NeurIPS 2020
Workshop paper

Entity and Relation Linking is All You Need for KGQA

Abstract

Ever since the advent of the Semantic Web, Knowledge Graph Question Answering (KGQA) has been a niche growing subfield within AI and NLP. The task of KGQA comprises answering a Natural Language Question from using the knowledge facts present in a given RDF based Knowledge Graphs (KGs) such as DBpedia, etc. The task of entity/relation linking is a crucial sub-step while designing any KGQA solution. There are many off-the-shelf entity/relation linkers available, for example Falcon but majority of the successful KGQA solutions do not go for such off-the-shelf entity/relations linkers because precision and recall of these linkers are not very high and instead follow the semantic parsing style approaches where they use their custom-designed linkers to translate the given question into an intermediate logic-form representation which is then converted into SPARQL and executed on KG to obtain the answer. In this work, we undertake a systematic study to find whether it's possible to design an effective modular approach comprising off-the-shelf linker and seq2seq style deep neural models. Our aim is to train a deep-net based model which can predict the SPARQL query and answer for a given test question.