
Research Article
SARA: Stochastic Adaption of Language Models to In-Domain RAG
@INPROCEEDINGS{10.4108/eai.21-11-2024.2354624, author={Peiyu Xu and Naxin Chen and Xinyu Liu and Denghao Peng}, title={SARA: Stochastic Adaption of Language Models to In-Domain RAG}, proceedings={Proceedings of the 2nd International Conference on Machine Learning and Automation, CONF-MLA 2024, November 21, 2024, Adana, Turkey}, publisher={EAI}, proceedings_a={CONF-MLA}, year={2025}, month={3}, keywords={stochastic adaption retrieval augmentation fine-tuning}, doi={10.4108/eai.21-11-2024.2354624} }
- Peiyu Xu
Naxin Chen
Xinyu Liu
Denghao Peng
Year: 2025
SARA: Stochastic Adaption of Language Models to In-Domain RAG
CONF-MLA
EAI
DOI: 10.4108/eai.21-11-2024.2354624
Abstract
In the current landscape of pre-trained large language model (LLM) applications, various methods such as data augmentation, Residual Learning (RL), Curriculum Learning (CL), Low-Rank Adaptation (LoRA) and Retrieval-Augmented Generation (RAG), are commonly employed to integrate more targeted information into pre-trained models. However, these methods often fall short in terms of training costs or practical effectiveness. Therefore, finding a more effective approach to enhance the capabilities of large language models in downstream tasks is imperative. In this paper, we introduce Stochastic Adaption of Retrieval Augmentation (SARA), an integrated fine-tuning strategy that introduces chained adaption stages while incorporating features from Retrieval-Augmented Fine-Tuning (RAFT), in order to enhance the LLM's ability of effectively responding to domain-specific queries. By organizing the various adaptation stages in a stochastic manner, we achieve improved performance, robustness and training efficiency compared to existing fine-tuning methods.