Welcome to the rLLM Project! 👋

We are a open-source initiative spawning from Berkeley Sky Computing Lab to democratize reinforcement learning (RL) techniques and develop scalable systems for large language models (LLMs) and agents.

rLLM v0.2: RL Training over General Agentic Programs

We are excited to announce rLLM v0.2, a major upgrade of our RL training framework. In v0.1, rLLM provided agent and OpenAI Gym-like environment abstractions to support training ReACT-style agents. In v0.2, we introduce AgentWorkflowEngine and AgentWorkflowTrainer—more general abstractions that enable arbitrary agentic programs to be trained. Agent builders and researchers can now define multi-agent systems, complex workflows (such as solver-judge, planner-executor, or MCTS), or agentic programs with custom reward functions, and train them with reinforcement learning without rewriting their production code.

Pepper: A Real-time, Event-Driven Architecture for Proactive Agentic Systems

We introduce Pepper, a real-time, event-driven architecture enabling the next generation of proactive agentic systems. With Pepper, we built a personal assistant that proactively fetches and summarizes emails, provides context before you even start a conversation, and continues to follow up while working on tasks. We open-source Pepper to advance the creation of proactive agentic systems.

rLLM: Reinforcement Learning for Language Agents

We release rLLM, an open-source framework for post-training language agents via reinforcement learning. With rLLM, you can easily build their custom agents and environments, train them with reinforcement learning, and deploy them for real-world workloads.

Project Leads

Members

Advisors