News • Links • Getting Started • Citation • Acknowledgement
SkyRL is a full-stack RL library that provides the following components:
skyagent
: Our agent layer for training long-horizon, real-world agents. Contains code for SkyRL-v0.- (NEW)
skyrl-train
: Our modular, performant training framework for RL. - (NEW)
skyrl-gym
: Our gynasium of tool-use tasks, including a library of math, coding, search and SQL environments implemented in the Gymnasium API.
For a guide on developing with SkyRL, take at look at our Development Guide docs.
For model training, checkout skyrl-train
to start using, modifying, or building on top of the SkyRL training stack. See our quickstart docs to ramp up!
For building environments, checkout skyrl-gym
to integrate your task in the simple gymnasium interface.
For agentic pipelines, checkout skyagent
for our work on optimizing and scaling pipelines for multi-turn tool use LLMs on long-horizon, real-environment tasks.
- [2025/06/26] 🎉 We released SkyRL-v0.1: A highly-modular, performant RL training framework. [Blog]
- [2025/06/26] 🎉 We released SkyRL-Gym: A library of RL environments for LLMs implemented with the Gymnasium API. [Blog]
- [2025/05/20] 🎉 We released SkyRL-SQL: a multi-turn RL training pipeline for Text-to-SQL, along with SkyRL-SQL-7B — a model trained on just 653 samples that outperforms both GPT-4o and o4-mini!
- [2025/05/06] 🎉 We released SkyRL-v0: our open RL training pipeline for multi-turn tool use LLMs, optimized for long-horizon, real-environment tasks like SWE-Bench!
This work is done at Berkeley Sky Computing Lab in collaboration with Anyscale, with generous compute support from Anyscale, Databricks, NVIDIA, Lambda Labs, and AMD.
We adopt many lessons and code from several great projects such as veRL, OpenRLHF, Search-R1, OpenReasonerZero, and NeMo-RL. We appreciate each of these teams and their contributions to open-source research!
If you find the work in this repository helpful, please consider citing:
@misc{cao2025skyrl,
title = {SkyRL-v0: Train Real-World Long-Horizon Agents via Reinforcement Learning},
author = {Shiyi Cao and Sumanth Hegde and Dacheng Li and Tyler Griggs and Shu Liu and Eric Tang and Jiayi Pan and Xingyao Wang and Akshay Malik and Graham Neubig and Kourosh Hakhamaneshi and Richard Liaw and Philipp Moritz and Matei Zaharia and Joseph E. Gonzalez and Ion Stoica},
year = {2025},
}
@misc{liu2025skyrlsql,
title={SkyRL-SQL: Matching GPT-4o and o4-mini on Text2SQL with Multi-Turn RL},
author={Shu Liu and Sumanth Hegde and Shiyi Cao and Alan Zhu and Dacheng Li and Tyler Griggs and Eric Tang and Akshay Malik and Kourosh Hakhamaneshi and Richard Liaw and Philipp Moritz and Matei Zaharia and Joseph E. Gonzalez and Ion Stoica},
year={2025},
}
@misc{griggs2025skrylv01,
title={Evolving SkyRL into a Highly-Modular RL Framework},
author={Tyler Griggs and Sumanth Hegde and Eric Tang and Shu Liu and Shiyi Cao and Dacheng Li and Charlie Ruan and Philipp Moritz and Kourosh Hakhamaneshi and Richard Liaw and Akshay Malik and Matei Zaharia and Joseph E. Gonzalez and Ion Stoica},
year={2025},
note={Notion Blog}
}