Multi-task retriever fine-tuning for domain-specific and efficient RAG
Abstract
Retrieval-Augmented Generation (RAG) has become ubiquitous when deploying Large Language Models (LLMs), as it can address typical limitations such as generating hallucinated or outdated information. However, when building real-world RAG applications, practical issues arise. First, the retrieved information is generally domain-specific. Since it is computationally expensive to fine-tune LLMs, it is more feasible to fine-tune the retriever to improve the quality of the data included in the LLM input. Second, as more applications are deployed in the same real-world system, one cannot afford to deploy separate retrievers. Moreover, these RAG applications normally retrieve different kinds of data. Our solution is to instruction fine-tune a small retriever encoder on a variety of domain-specific tasks to allow us to deploy one encoder that can serve many use cases, thereby achieving low-cost, scalability, and speed. We show how this encoder generalizes to out-of-domain settings as well as to an unseen retrieval task on real-world enterprise use cases.
Community
We present a step-by-step guide on how to finetune retriever models for domain-specific RAG. If you have questions about our paper, send it my way!
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- CodeXEmbed: A Generalist Embedding Model Family for Multiligual and Multi-task Code Retrieval (2024)
- Large Language Model Can Be a Foundation for Hidden Rationale-Based Retrieval (2024)
- Invar-RAG: Invariant LLM-aligned Retrieval for Better Generation (2024)
- LLMs are Also Effective Embedding Models: An In-depth Overview (2024)
- Auto-RAG: Autonomous Retrieval-Augmented Generation for Large Language Models (2024)
- Towards Optimizing a Retrieval Augmented Generation using Large Language Model on Academic Data (2024)
- GME: Improving Universal Multimodal Retrieval by Multimodal LLMs (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper