@inproceedings{4c9a13a7e7254adeb81c9f84e5f94fce,
title = "Knowledge Base Construction from Pre-trained Language Models by Prompt learning",
abstract = "Pre-trained language models (LMs) have advanced the state-of-the-art for many semantic tasks and have also been proven effective for extracting knowledge from the models itself. Although several works have explored the capability of the LMs for constructing knowledge bases, including prompt learning, this potential has not yet been fully explored. In this work, we propose a method of extracting factual knowledge from LMs for given subject-relation pairs and explore the most effective strategy to generate blank object entities for each relation of triples. We design prompt templates for each relation using personal knowledge and the descriptive information available on the web such as WikiData. The probing approach of our proposed LMs is tested on the dataset provided by the International Semantic Web Conference (ISWC 2022) LM-KBC Challenge. To cope with the problem of varying performance for each relation, we designed a parameter selection strategy for each relation. Using the test dataset, we obtain an F1-score of 0.4935%, which is higher than the baseline of 31.08%.",
keywords = "Information Extraction, Link Prediction, Pre-trained language model, Prompt learning",
author = "Xiao Ning and Remzi Celebi",
note = "Funding Information: Thanks to Shuai Wang, an excellent software engineer from Amazon, he introduced several practical scripts for me to automate run the code, which significantly increased the efficiency of experiments. Furthermore, the experiment part of this research was made possible, in part, using the Data Science Research Infrastructure (DSRI) hosted at Maastricht University. Publisher Copyright: {\textcopyright} 2022 Copyright for this paper by its authors.; 2022 Semantic Web Challenge on Knowledge Base Construction from Pre-Trained Language Models, LM-KBC 2022 ; Conference date: 01-01-2022 Through 01-10-2022",
year = "2022",
month = jan,
day = "1",
language = "English",
volume = "3274",
series = "CEUR Workshop Proceedings",
publisher = "Rheinisch-Westfaelische Technische Hochschule Aachen * Lehrstuhl Informatik V",
pages = "46--54",
booktitle = "Knowledge Base Construction from Pre-trained Language Models 2022",
}