上下文語言模型化技術於常見問答檢索之研究

Translated title of the contribution: A Study on Contextualized Language Modeling for FAQ Retrieval

Wen Ting Tseng, Yung Chang Hsu, Berlin Chen

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Recent years have witnessed significant progress in the development of deep learning techniques, which also has achieved state-of-the-art performance for a wide variety of natural language processing (NLP) applications like the frequently asked question (FAQ) retrieval task. FAQ retrieval, which manages to provide relevant information in response to frequent questions or concerns, has far-reaching applications such as e-commerce services and online forums, among many other applications. In the common setting of the FAQ retrieval task, a collection of question-answer (Q-A) pairs compiled in advance can be capitalized to retrieve an appropriate answer in response to a user’s query that is likely to reoccur frequently. To date, there have many strategies proposed to approach FAQ retrieval, ranging from comparing the similarity between the query and a question, to scoring the relevance between the query and the associated answer of a question, and performing classification on user queries. As such, a variety of contextualized language models have been extended and developed to operationalize the aforementioned strategies, like BERT (Bidirectional Encoder Representations from Transformers), K-BERT and Sentence-BERT. Although BERT and its variants has demonstrated reasonably good results on various FAQ retrieval tasks, they still would fall short for some tasks that may resort to generic knowledge. In view of this, in this paper, we set out to explore the utility of injecting an extra knowledge base into BERT for FAQ retrieval, meanwhile comparing among synergistic effects of different strategies and methods.

Translated title of the contributionA Study on Contextualized Language Modeling for FAQ Retrieval
Original languageChinese (Traditional)
Title of host publicationROCLING 2020 - 32nd Conference on Computational Linguistics and Speech Processing
EditorsJenq-Haur Wang, Ying-Hui Lai, Lung-Hao Lee, Kuan-Yu Chen, Hung-Yi Lee, Chi-Chun Lee, Syu-Siang Wang, Hen-Hsen Huang, Chuan-Ming Liu
PublisherThe Association for Computational Linguistics and Chinese Language Processing (ACLCLP)
Pages247-259
Number of pages13
ISBN (Electronic)9789869576932
Publication statusPublished - 2020
Event32nd Conference on Computational Linguistics and Speech Processing, ROCLING 2020 - Taipei, Taiwan
Duration: 2020 Sept 242020 Sept 26

Publication series

NameROCLING 2020 - 32nd Conference on Computational Linguistics and Speech Processing

Conference

Conference32nd Conference on Computational Linguistics and Speech Processing, ROCLING 2020
Country/TerritoryTaiwan
CityTaipei
Period2020/09/242020/09/26

ASJC Scopus subject areas

  • Language and Linguistics
  • Speech and Hearing

Fingerprint

Dive into the research topics of 'A Study on Contextualized Language Modeling for FAQ Retrieval'. Together they form a unique fingerprint.

Cite this