On the Complementary Nature of Knowledge Graph Embedding, Fine Grain Entity Types, and Language Modeling
We demonstrate the complementary natures of neural knowledge graph embedding, fine-grain entity type prediction, and neural language modeling. We show that a language model-inspired knowledge graph embedding approach yields both improved knowledge graph embeddings and fine-grain entity type representations. Our work also shows that jointly modeling both structured knowledge tuples and language improves both.
PDF Abstract EMNLP (DeeLIO) 2020 PDF EMNLP (DeeLIO) 2020 AbstractDatasets
Results from the Paper
Submit
results from this paper
to get state-of-the-art GitHub badges and help the
community compare results to other papers.