[2025/08/14] Urgent
[2025/07/14] Make the repository open source!
AdRE is a LoRA Adapter reinforcement learning (RL) framework designed for large language models (LLMs) and supports multi-expert (Multi-LoRA-Expert) architecture. It supports reinforcement learning training (such as GRPO) and Adapter-based efficient fine-tuning (SFT). This framework is suitable for research and practical application scenarios such as RLHF and efficient LLM adaptation.
Multi-LoRA Architecture
- Supports a Multi-LoRA architecture with three components:
- Frozen layers: not trainable.
- Routing layers: dynamic MoE-style routing based on attention.
- Multi-LoRA layers: trainable layers composed of multiple LoRA experts.
LoRA SFT (Supervised Fine-tuning)
- Enables efficient multi-GPU fine-tuning for LoRA and Multi-LoRA using data parallelism.
Reinforcement Learning (RL)
- Supports LoRA-based RL fine-tuning based on GRPO, with multi-GPU data parallelism.
- Built-in support for:
- Reward modeling
- Experience collection
- Temperature scheduling
- Requires only a single model with LoRA adapters (no separate reference model needed), significantly reducing GPU memory usage.
- Supports layer freezing and shared computation between the reference and actor models to save resources.
@software{adre_2025,
author = {Mengqi Liao},
title = {AdRE: Adapter Reinforcement Learning Framework},
url = {https://github.com/LiaoMengqi/AdRE},
version = {1.0.0},
date = {2025-07-14},
license = {MIT},
publisher = {GitHub},
howpublished = {GitHub repository}
}
This project is inspired by and references the excellent work of OpenRLHF and VeRL. Sincere thanks to their open-source contributions to the community.
