Welcome to Trinity-RFTβs documentation!#
π‘ What is Trinity-RFT?#
Trinity-RFT is a general-purpose, flexible and user-friendly framework for LLM reinforcement fine-tuning (RFT). It decouples RFT into three components that work in coordination:
Explorer generates experience data via agent-environment interaction;
Trainer updates model weights by minimizing losses on the data;
Buffer pipelines data processing throughout the RFT lifecycle.
Trinity-RFT provides functionalities for users with different backgrounds and objectives:
π€ Agent application developers: Train LLM-powered agents and improve their capabilities in specific domains [tutorial]
π§ Reinforcement learning researchers: Design, implement and validate new RL algorithms using compact, plug-and-play modules that allow non-invasive customization [tutorial]
π Data engineers: Create RFT datasets and build data pipelines for cleaning, augmentation, and human-in-the-loop scenarios [tutorial]
π¨ Tutorials and Guidelines#
Category |
Tutorial / Guideline |
|---|---|
Run diverse RFT modes |
+ Quick start: GRPO on GSM8k |
Multi-step agentic RL |
+ Concatenated multi-turn workflow |
Full-lifecycle data pipelines |
+ Rollout task mixing and selection |
Algorithm development |
+ RL algorithm development with Trinity-RFT (π paper) |
Going deeper into Trinity-RFT |
+ Full configurations |
π Key Features#
Flexible RFT Modes:
Supports synchronous/asynchronous, on-policy/off-policy, and online/offline RL.
Rollout and training can run separately and scale independently across devices.
Boost sample and time efficiency by experience replay.
Agentic RL Support:
Supports both concatenated and general multi-step agentic workflows.
Able to directly train agent applications developed using agent frameworks like AgentScope.
Full-Lifecycle Data Pipelines:
Enables pipeline processing of rollout tasks and experience samples.
Active data management (prioritization, cleaning, augmentation, etc.) throughout the RFT lifecycle.
Native support for multi-task joint learning and online task curriculum construction.
User-Friendly Design:
Plug-and-play modules and decoupled architecture, facilitating easy adoption and development.
Rich graphical user interfaces enable low-code usage.
Acknowledgements#
This project is built upon many excellent open-source projects, including:
verl and PyTorchβs FSDP for LLM training;
vLLM for LLM inference;
Data-Juicer for data processing pipelines;
AgentScope for agentic workflow;
Ray for distributed systems;
we have also drawn inspirations from RL frameworks like OpenRLHF, TRL and ChatLearn;
β¦β¦
Citation#
@misc{trinity-rft,
title={Trinity-RFT: A General-Purpose and Unified Framework for Reinforcement Fine-Tuning of Large Language Models},
author={Xuchen Pan and Yanxi Chen and Yushuo Chen and Yuchang Sun and Daoyuan Chen and Wenhao Zhang and Yuexiang Xie and Yilun Huang and Yilei Zhang and Dawei Gao and Yaliang Li and Bolin Ding and Jingren Zhou},
year={2025},
eprint={2505.17826},
archivePrefix={arXiv},
primaryClass={cs.LG},
url={https://arxiv.org/abs/2505.17826},
}