TY - GEN
T1 - KNOT-MCTS
T2 - 35th Conference on Computational Linguistics and Speech Processing, ROCLING 2023
AU - Wu, Chung Wen
AU - Huang, Guan Tang
AU - He, Yue Yang
AU - Chen, Berlin
N1 - Publisher Copyright:
© 2023 ROCLING 2023 - Proceedings of the 35th Conference on Computational Linguistics and Speech Processing. All rights reserved.
PY - 2023
Y1 - 2023
N2 - Contemporary large language models (LLMs) have made significant advancements, capable of generating fluent conversations with humans and accomplishing various tasks such as programming and question answering (QA). Nevertheless, current LLMs are still faced with numerous challenges, including generating hallucinations, lacking the latest information, suffering from biases, and others. In this paper, we proposed a technique, Knowledge-based Navigation for Optimal Truthfulness Monte Carlo Tree Search (KNOT-MCTS), which can reduce hallucinations of LLMs by aligning semantics of responses with external knowledge during the generation process. This technique acts as a plug-and-play knowledge injection method, which does not require any training and can be applied to any (large) language model. First, we retrieve relevance knowledge snippets, incorporating them into the prompt section and subsequently fed into the decoding process. Then, during the decoding process, we utilize our semantic alignment heuristic function to guide the response generation process of LMs through the Monte Carlo Tree Search (MCTS) decoding process. In our experiments on the TruthfulQA dataset, KNOT-MCTS paired with various LMs consistently outperforms their respective baselines. Our results demonstrate that KNOT-MCTS can effectively inject knowledge into various LMs to reduce hallucinations of LMs.
AB - Contemporary large language models (LLMs) have made significant advancements, capable of generating fluent conversations with humans and accomplishing various tasks such as programming and question answering (QA). Nevertheless, current LLMs are still faced with numerous challenges, including generating hallucinations, lacking the latest information, suffering from biases, and others. In this paper, we proposed a technique, Knowledge-based Navigation for Optimal Truthfulness Monte Carlo Tree Search (KNOT-MCTS), which can reduce hallucinations of LLMs by aligning semantics of responses with external knowledge during the generation process. This technique acts as a plug-and-play knowledge injection method, which does not require any training and can be applied to any (large) language model. First, we retrieve relevance knowledge snippets, incorporating them into the prompt section and subsequently fed into the decoding process. Then, during the decoding process, we utilize our semantic alignment heuristic function to guide the response generation process of LMs through the Monte Carlo Tree Search (MCTS) decoding process. In our experiments on the TruthfulQA dataset, KNOT-MCTS paired with various LMs consistently outperforms their respective baselines. Our results demonstrate that KNOT-MCTS can effectively inject knowledge into various LMs to reduce hallucinations of LMs.
KW - Knowledge Injection
KW - Knowledge Retrieval
KW - Monte Carlo Tree Search
KW - Semantic Alignment
UR - http://www.scopus.com/inward/record.url?scp=85184843521&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85184843521&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85184843521
T3 - ROCLING 2023 - Proceedings of the 35th Conference on Computational Linguistics and Speech Processing
SP - 215
EP - 221
BT - ROCLING 2023 - Proceedings of the 35th Conference on Computational Linguistics and Speech Processing
A2 - Wu, Jheng-Long
A2 - Su, Ming-Hsiang
A2 - Huang, Hen-Hsen
A2 - Tsao, Yu
A2 - Tseng, Hou-Chiang
A2 - Chang, Chia-Hui
A2 - Lee, Lung-Hao
A2 - Liao, Yuan-Fu
A2 - Ma, Wei-Yun
PB - The Association for Computational Linguistics and Chinese Language Processing (ACLCLP)
Y2 - 20 October 2023 through 21 October 2023
ER -