Loading [MathJax]/extensions/MathMenu.js
Utilize Pre-Trained PhoBERT to Compute Text Similarity and Rerank Documents for Question-Answering Task | IEEE Conference Publication | IEEE Xplore

Utilize Pre-Trained PhoBERT to Compute Text Similarity and Rerank Documents for Question-Answering Task


Abstract:

Open-domain Question Answering (QA) is a crucial task in natural language processing. QA systems typically follow two main steps: (i) identifying relevant passages and (i...Show More

Abstract:

Open-domain Question Answering (QA) is a crucial task in natural language processing. QA systems typically follow two main steps: (i) identifying relevant passages and (ii) generating answer sentences from these passages. Among these steps, identifying relevant passages poses a greater challenge and requires further refinement. In this paper, we introduce two novel strategies to improve the performance of this step, including: (i) a new method for computing the similarity between questions and text passages, and (ii) the integration of pretrained and fine-tuned models. Empirical evaluations conducted on the Zalo 2022 dataset demonstrate the efficacy of our proposed methods, manifesting a notable 10% increase in recall compared to using the BM25 method alone, and a 6% increase in recall compared to relying solely on a fine-tuned cross-encoder model.
Date of Conference: 27-29 November 2023
Date Added to IEEE Xplore: 10 January 2024
ISBN Information:

ISSN Information:

Conference Location: Hanoi, Vietnam

Contact IEEE to Subscribe

References

References is not available for this document.