TY - GEN
T1 - A Study on Contextualized Language Modeling for Machine Reading Comprehension
AU - Wu, Chin Ying
AU - Hsu, Yung Chang
AU - Chen, Berlin
N1 - Publisher Copyright:
© 2021 ROCLING 2021 - Proceedings of the 33rd Conference on Computational Linguistics and Speech Processing. All rights reserved.
PY - 2021
Y1 - 2021
N2 - With the recent breakthrough of deep learning technologies, research on machine reading comprehension (MRC) has attracted much attention and found its versatile applications in many use cases. MRC is an important natural language processing (NLP) task aiming to assess the ability of a machine to understand natural language expressions, which is typically operationalized by first asking questions based on a given text paragraph and then receiving machine-generated answers in accordance with the given context paragraph and questions. In this paper, we leverage two novel pretrained language models built on top of Bidirectional Encoder Representations from Transformers (BERT), namely BERT-wwm and MacBERT, to develop effective MRC methods. In addition, we also seek to investigate whether additional incorporation of the categorical information about a context paragraph can benefit MRC or not, which is achieved based on performing context paragraph clustering on the training dataset. On the other hand, an ensemble learning approach is proposed to harness the synergistic power of the aforementioned two BERT-based models so as to furtherpromoteMRCperformance.
AB - With the recent breakthrough of deep learning technologies, research on machine reading comprehension (MRC) has attracted much attention and found its versatile applications in many use cases. MRC is an important natural language processing (NLP) task aiming to assess the ability of a machine to understand natural language expressions, which is typically operationalized by first asking questions based on a given text paragraph and then receiving machine-generated answers in accordance with the given context paragraph and questions. In this paper, we leverage two novel pretrained language models built on top of Bidirectional Encoder Representations from Transformers (BERT), namely BERT-wwm and MacBERT, to develop effective MRC methods. In addition, we also seek to investigate whether additional incorporation of the categorical information about a context paragraph can benefit MRC or not, which is achieved based on performing context paragraph clustering on the training dataset. On the other hand, an ensemble learning approach is proposed to harness the synergistic power of the aforementioned two BERT-based models so as to furtherpromoteMRCperformance.
KW - Deep Learning
KW - Language model
KW - Machine Reading Comprehension
KW - Natural Language Processing
UR - http://www.scopus.com/inward/record.url?scp=85127376396&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85127376396&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85127376396
T3 - ROCLING 2021 - Proceedings of the 33rd Conference on Computational Linguistics and Speech Processing
SP - 48
EP - 57
BT - ROCLING 2021 - Proceedings of the 33rd Conference on Computational Linguistics and Speech Processing
A2 - Lee, Lung-Hao
A2 - Chang, Chia-Hui
A2 - Chen, Kuan-Yu
PB - The Association for Computational Linguistics and Chinese Language Processing (ACLCLP)
T2 - 33rd Conference on Computational Linguistics and Speech Processing, ROCLING 2021
Y2 - 15 October 2021 through 16 October 2021
ER -