Benchmarking Docs License Python

AgentJet

AgentJet (AJet) is a cutting-edge, user-friendly agent RL training framework designed to optimize agents and agentic workflows (supporting any agent built with OpenAI SDK, AgentScope, Langchain, or raw HTTP requests), fine-tuning LLM weights to enhance model performance.

AgentJet (AJet) has fully-distributed swarm training capability, which means that you can deploy ajet-swarm start in GPU server(s) and then start training agents in your laptop(s)! Simply provide your agent workflow, training dataset, and reward function, and AgentJet will be ready to go!

Fast Introduction

Classic Mode

Let's begin with the simplest example: a math agent with a tool call. This is a simple & centralized training example.

  1. please check out the installation guide to set up the training environment.
  2. tune your first model using the minimum example.
    ajet --conf ./tutorial/example_math_agent/math_agent.yaml --backbone='verl'
    
image

Swarm Mode

Let's begin with the simplest AgentJet Swarm example: also a math agent. In this case, you can use any GPU-less laptop to train the model remotely.

  1. Start swarm server and begin swarm overwatch: ajet-swarm start and ajet-swarm overwatch. (Alternative: if you are a fan of docker, use our prebuilt docker image here without setting up dependencies)
  2. From your laptop (or swarm server localhost), run this simple script to begin training:
    AJET_SWARM_URL="http://swarm-server-ip:10086" python ./tutorial/example_math_swarm/math.py
    
image

Key Features

image

Example Library

Explore our rich library of examples to kickstart your journey:

image

Project Structure

AgentJet makes agent fine-tuning straightforward by separating the developer interface from the internal execution logic.

Basic Modules

To optimize an agent, you provide three core inputs:

AgentJet Architecture

The internal system orchestrates several specialized modules to handle the complexities of RL training and agent interactions.

Module Description
Launcher Manages background service processes (Ray, vLLM) and routes the backbone
Task Rollout Bridges LLM engines and manages the Gym environment lifecycle
Task Runner Executes the agent workflow and calculates rewards
Model Tuner Forwards inference requests from the workflow to the LLM engine
Context Tracker Monitors LLM calls and automatically merges shared-history timelines (1.5x-10x efficiency boost)
Swarm Server A data interchange center that accepts OpenAI-like requests and engine instructions, activated only in AgentJet Swarm mode

Swarm Architecture

When swarm training mode is enabled, an additional component will be activated:

  • Swarm Data Interchange Server: Maintains HTTP service, listens to swarm instructions and OpenAI compatible requests. Establishes a high-speed zmq communication channel to coordinate other modules.
image

Roadmap

AgentJet is a constantly evolving project. We are planning to add the following features in the near future.

Category Feature Status
Examples Add LoRA training examples Todo
Infra Optimize configurations for long-context adaptation on smaller GPUs In Progress
Capability Multi-modal training support Todo
Capability MARL Credit assignment Todo
Capability Training dataset generation from few-shot samples Todo

Citation

If you use AgentJet in your research, please cite:

@software{
  title  = {AgentJet: A Cutting-Edge Multi-Agent Training Platform for Large Language Models.},
  author = {The AgentJet Team},
  url    = {https://modelscope.github.io/AgentJet/},
  month  = {01},
  year   = {2026}
}

Next Steps


This project is under active development, we need your help to make it shine!
[⭐ Star Us](https://github.com/modelscope/AgentJet) · [Report Bug](https://github.com/modelscope/AgentJet/issues) · [Request Feature](https://github.com/modelscope/AgentJet/issues)
image
Join AgentJet DingTalk Group to share your idea