TY - GEN
T1 - Knowledge Base Construction from Pre-trained Language Models by Prompt learning
AU - Ning, Xiao
AU - Celebi, Remzi
N1 - Funding Information:
Thanks to Shuai Wang, an excellent software engineer from Amazon, he introduced several practical scripts for me to automate run the code, which significantly increased the efficiency of experiments. Furthermore, the experiment part of this research was made possible, in part, using the Data Science Research Infrastructure (DSRI) hosted at Maastricht University.
Publisher Copyright:
© 2022 Copyright for this paper by its authors.
PY - 2022/1/1
Y1 - 2022/1/1
N2 - Pre-trained language models (LMs) have advanced the state-of-the-art for many semantic tasks and have also been proven effective for extracting knowledge from the models itself. Although several works have explored the capability of the LMs for constructing knowledge bases, including prompt learning, this potential has not yet been fully explored. In this work, we propose a method of extracting factual knowledge from LMs for given subject-relation pairs and explore the most effective strategy to generate blank object entities for each relation of triples. We design prompt templates for each relation using personal knowledge and the descriptive information available on the web such as WikiData. The probing approach of our proposed LMs is tested on the dataset provided by the International Semantic Web Conference (ISWC 2022) LM-KBC Challenge. To cope with the problem of varying performance for each relation, we designed a parameter selection strategy for each relation. Using the test dataset, we obtain an F1-score of 0.4935%, which is higher than the baseline of 31.08%.
AB - Pre-trained language models (LMs) have advanced the state-of-the-art for many semantic tasks and have also been proven effective for extracting knowledge from the models itself. Although several works have explored the capability of the LMs for constructing knowledge bases, including prompt learning, this potential has not yet been fully explored. In this work, we propose a method of extracting factual knowledge from LMs for given subject-relation pairs and explore the most effective strategy to generate blank object entities for each relation of triples. We design prompt templates for each relation using personal knowledge and the descriptive information available on the web such as WikiData. The probing approach of our proposed LMs is tested on the dataset provided by the International Semantic Web Conference (ISWC 2022) LM-KBC Challenge. To cope with the problem of varying performance for each relation, we designed a parameter selection strategy for each relation. Using the test dataset, we obtain an F1-score of 0.4935%, which is higher than the baseline of 31.08%.
KW - Information Extraction
KW - Link Prediction
KW - Pre-trained language model
KW - Prompt learning
M3 - Conference article in proceeding
VL - 3274
T3 - CEUR Workshop Proceedings
SP - 46
EP - 54
BT - Knowledge Base Construction from Pre-trained Language Models 2022
T2 - 2022 Semantic Web Challenge on Knowledge Base Construction from Pre-Trained Language Models
Y2 - 1 January 2022 through 1 October 2022
ER -