About | Contact Us | Register | Login
ProceedingsSeriesJournalsSearchEAI
Proceedings of the 2nd International Conference on Machine Learning and Automation, CONF-MLA 2024, November 21, 2024, Adana, Turkey

Research Article

SARA: Stochastic Adaption of Language Models to In-Domain RAG

Download93 downloads
Cite
BibTeX Plain Text
  • @INPROCEEDINGS{10.4108/eai.21-11-2024.2354624,
        author={Peiyu  Xu and Naxin  Chen and Xinyu  Liu and Denghao  Peng},
        title={SARA: Stochastic Adaption of Language Models to In-Domain RAG},
        proceedings={Proceedings of the 2nd International Conference on Machine Learning and Automation, CONF-MLA 2024, November 21, 2024, Adana, Turkey},
        publisher={EAI},
        proceedings_a={CONF-MLA},
        year={2025},
        month={3},
        keywords={stochastic adaption retrieval augmentation fine-tuning},
        doi={10.4108/eai.21-11-2024.2354624}
    }
    
  • Peiyu Xu
    Naxin Chen
    Xinyu Liu
    Denghao Peng
    Year: 2025
    SARA: Stochastic Adaption of Language Models to In-Domain RAG
    CONF-MLA
    EAI
    DOI: 10.4108/eai.21-11-2024.2354624
Peiyu Xu1,*, Naxin Chen2, Xinyu Liu3, Denghao Peng4
  • 1: UM-SJTU Joint Institute, Shanghai Jiao Tong University
  • 2: Central China Normal University
  • 3: Jilin University
  • 4: University of Macau
*Contact email: xupeiyu0827@sjtu.edu.cn

Abstract

In the current landscape of pre-trained large language model (LLM) applications, various methods such as data augmentation, Residual Learning (RL), Curriculum Learning (CL), Low-Rank Adaptation (LoRA) and Retrieval-Augmented Generation (RAG), are commonly employed to integrate more targeted information into pre-trained models. However, these methods often fall short in terms of training costs or practical effectiveness. Therefore, finding a more effective approach to enhance the capabilities of large language models in downstream tasks is imperative. In this paper, we introduce Stochastic Adaption of Retrieval Augmentation (SARA), an integrated fine-tuning strategy that introduces chained adaption stages while incorporating features from Retrieval-Augmented Fine-Tuning (RAFT), in order to enhance the LLM's ability of effectively responding to domain-specific queries. By organizing the various adaptation stages in a stochastic manner, we achieve improved performance, robustness and training efficiency compared to existing fine-tuning methods.

Keywords
stochastic adaption retrieval augmentation fine-tuning
Published
2025-03-11
Publisher
EAI
http://dx.doi.org/10.4108/eai.21-11-2024.2354624
Copyright © 2024–2025 EAI
EBSCOProQuestDBLPDOAJPortico
EAI Logo

About EAI

  • Who We Are
  • Leadership
  • Research Areas
  • Partners
  • Media Center

Community

  • Membership
  • Conference
  • Recognition
  • Sponsor Us

Publish with EAI

  • Publishing
  • Journals
  • Proceedings
  • Books
  • EUDL