Scale up the large language models to store vast amounts of knowledge within their parameters incur higher costs and training times. Thus, in this study, we aim to examine the effects of language models enhancing external knowledge and compare the performance of extractive and abstractive generation tasks in building the question-answering system. To ensure consistency in our evaluations, we modified the MS MARCO and MASH-QA datasets by filtering irrelevant support documents and enhancing contextual relevance by mapping the input question to the closest supported documents in our database setup. Finally, we materiality assess the performance in the health domain, our experience presents a promising result not only with information retrieval but also with retrieval augmentation tasks aimed at improving performance for future work.
发表评论