Enhancing LLM Question Answering with RAG through Dense Vector Search and Re-Ranking

  • Te Lun Yang*
  • , Jyi Shane Liu
  • , Yuen Hsien Tseng
  • , Jyh Shing Jang
  • , Ming Ching Chang
  • , Wei Chao Chen
  • *Corresponding author for this work

Research output: Contribution to journalConference articlepeer-review

Abstract

Retrieval-Augmented Generation (RAG) has emerged as a powerful framework for enhancing Large Language Models (LLMs) by incorporating external knowledge through information retrieval (IR) techniques. However, in question-answering tasks, RAG often retrieves documents that are only semantically similar to the query, which may not provide the most relevant information for generating accurate responses. To address this limitation, we propose an improved retrieval pipeline that combines dense vector search with a re-ranking mechanism to more effectively identify and extract highly relevant knowledge from the retrieved content. We evaluated our approach on two Chinese datasets, TTQA and TMMLU+, using 17 different LLMs. Experimental results show that our method improves performance by up to 21.24% over baseline approaches, particularly on two finance-related subsets, after incorporating domain-specific financial regulations to enhance the knowledge base used in the TMMLU+ dataset.

ASJC Scopus subject areas

  • Signal Processing
  • Media Technology
  • Computer Vision and Pattern Recognition
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'Enhancing LLM Question Answering with RAG through Dense Vector Search and Re-Ranking'. Together they form a unique fingerprint.

Cite this