Skip to content

alibaba/ROLL

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

ROLL Logo

ROLL: Reinforcement Learning Optimization for Large-Scale Learning

๐Ÿš€ An Efficient and User-Friendly Scaling Library for Reinforcement Learning with Large Language Models ๐Ÿš€

License GitHub issues Repo stars WeChat QR

ROLL is an efficient and user-friendly RL library designed for Large Language Models (LLMs) utilizing Large Scale GPU resources. It significantly enhances LLM performance in key areas such as human preference alignment, complex reasoning, and multi-turn agentic interaction scenarios.

Leveraging a multi-role distributed architecture with Ray for flexible resource allocation and heterogeneous task scheduling, ROLL integrates cutting-edge technologies like Megatron-Core, SGLang and vLLM to accelerate model training and inference.


๐Ÿ“ข News

๐Ÿ“ฃ Updates
[08/11/2025] ๐ŸŽ‰ Our Paper released, see Part I: Tricks or Traps? A Deep Dive into RL for LLM Reasoning.
[08/13/2025] ๐ŸŽ‰ ROLL supports AMD GPUs with out-of-box image docker and Dockerfile and specific yamls under examples/ directory. Please refer to Installation.
[08/10/2025] ๐ŸŽ‰ Agentic RL supports stepwise learning, like GiGPO; Distill supports VLM. Explore the new capabilities!
[07/31/2025] ๐ŸŽ‰ Refactor agentic rl design. Support agentic rl async training. Explore the new capabilities!
[07/31/2025] ๐ŸŽ‰ Support DistillPipeline/DpoPipeline. Support lora. Support GSPO
[06/25/2025] ๐ŸŽ‰ Support thread env for env scaling and support qwen2.5 VL agentic pipeline.
[06/13/2025] ๐ŸŽ‰ Support Qwen2.5 VL rlvr pipeline and upgrade mcore to 0.12 version.
[06/09/2025] ๐ŸŽ‰ ROLL tech report is now available! Access the report here.
[05/30/2025] ๐ŸŽ‰ Training RLVR and Agentic RL with ROLL is now available! Explore the new capabilities.

๐Ÿš€ Get Started

Documents

Quick Start

Quick Start based on alicloud
Installation
Config guide

Step By Step

RLVR Pipeline
Agentic RL Pipeline


โœจ Key Features

  • Multi-task RL Training (RLVR): Covers mathematics, coding, general reasoning, open-ended Q&A, instruction following, etc.
    • Flexible domain_batch_size distribution control.
    • Sample-level asynchronous parallel Rollout, asynchronous reward calculation, and dynamic sampling.
    • Asynchronous training under implementation.
  • Agentic RL: Multi-turn interaction capabilities for games, multi-turn dialogues, tool use, etc.
    • Environment-level asynchronous parallel rollout.
    • Supports asynchronous training.
    • Multi-turn interaction rollout supports local debugging, improving multi-turn interaction business development efficiency.
    • Supports TrajectoryWise (StartPO) and StepWise (GiGPO) training paradigms.
  • Algorithm-Friendly: Provides flexible and rich RL strategy configurations by default.
    • Over 20 rich reinforcement learning strategy options, such as reward normalization, reward clipping, various advantage estimation methods, etc.
    • Out-of-the-box support for reinforcement learning algorithms, such as PPO, GRPO, Reinforce++, TOPR, RAFT++, GSPO, etc.
  • Rich Training and Inference Engine: Ray-based multi-role distributed architecture; Strategy abstraction unifies various backends, enabling easy operation from single machines to thousands-of-GPU clusters.
    • Inference/Generation supports vLLM, SGLang.
    • Training supports DeepSpeed (ZeRO), Megatron-LM 5D parallelism (mcore-adapter, dp/tp/pp/cp/ep), FSDP under implementation.
    • Extreme offload/reload capabilities.
    • Supports LoRA training.
    • Supports FP8 rollout (FP8 inference for LLM as judge, FP8 rollout with BF16 training under development).
  • AutoDeviceMapping: Supports custom device mapping for different roles, flexibly managing colocated and disaggregated deployments.
  • Observability: Integrated with SwanLab / WandB / TensorBoard, tracking of performance for each domain and reward type.
  • Rich Post-training Technical Support:
    • Agentic RL LLM & VLM
    • RLVR LLM & VLM
    • Distill Pipeline LLM & VLM
    • DPO Pipeline
    • SFT Pipeline under development

๐Ÿ”ฎ Upcoming Features

We are continuously working to expand ROLL's capabilities:

  • โฑ๏ธ Async RLVR pipeline: For even more efficient and streamlined asynchronous operations.
  • โš™๏ธ FSDP2: Integrating the latest Fully Sharded Data Parallel techniques.
  • ๐Ÿ” Support DeepseekV3: Adding compatibility for the newest Deepseek models.

๐Ÿ† Notable work based on ROLL

  • RecGPT: a next-generation, LLM-driven framework that places user intent at the core of recommender systems, fostering a more sustainable and mutually beneficial ecosystem.

๐Ÿ™ Citation and Acknowledgement

ROLL is inspired by the design of OpenRLHF, VeRL, Nemo-Aligner, and RAGEN. The project is developed by Alibaba TAOBAO & TMALL Group and Alibaba Group. The code is distributed under the Apache License (Version 2.0). This product contains various third-party components under other open-source licenses. See the NOTICE file for more information.

The following repositories have been used in ROLL, either in their close-to-original form or as an inspiration:

If you use ROLL in your research or project, please consider citing us:

@article{wang2025reinforcement,
  title={Reinforcement Learning Optimization for Large-Scale Learning: An Efficient and User-Friendly Scaling Library},
  author={Wang, Weixun and Xiong, Shaopan and Chen, Gengru and Gao, Wei and Guo, Sheng and He, Yancheng and Huang, Ju and Liu, Jiaheng and Li, Zhendong and Li, Xiaoyang and others},
  journal={arXiv preprint arXiv:2506.06122},
  year={2025}
}

๐Ÿค About [ROLL Team]

ROLL is a project jointly developed by Taotian Future Life Lab and Aicheng Technology, with a strong emphasis on pioneering the future of Reinforcement Learning (RL). Our mission is to explore and shape innovative forms of future living powered by advanced RL technologies. If you are passionate about the future of RL and want to be part of its evolution, we warmly welcome you to join us! Learn more about the ROLL Team through our official channels below๐Ÿ‘‡

WeChat QR

We are HIRING!

  • Post Training Infra ็ ”ๅ‘ๅทฅ็จ‹ๅธˆ JD link
  • ๅคงๆจกๅž‹่ฎญ็ปƒไธ“ๅฎถ๏ผš
    • ๏ผˆ็คพๆ‹›๏ผ‰JD link
    • ๏ผˆๆ กๆ‹›๏ผ‰JD link
  • Infra ็ ”็ฉถๅž‹ๅฎžไน ็”Ÿ JD link

We welcome contributions from the community! ๐Ÿค

About

An Efficient and User-Friendly Scaling Library for Reinforcement Learning with Large Language Models

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 31