Skip to content

Commit 77835ab

Browse files
committed
Refactor AgentJet Swarm Documentation and Architecture
- Deleted outdated swarm blueprint and main documentation files. - Added new detailed Swarm Design Blueprint and overview documentation in English. - Updated mkdocs configuration to reflect new documentation structure. - Introduced a comprehensive blog post on the AgentJet Swarm framework, highlighting its architecture, advantages, and use cases for distributed multi-agent LLM training.
1 parent 504de1d commit 77835ab

File tree

5 files changed

+170
-18
lines changed

5 files changed

+170
-18
lines changed

README.md

Lines changed: 17 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@
1818

1919

2020

21-
## ✈️ Minimum Example
21+
## ✈️ Fast Introduction
2222

2323
### Classic Mode
2424

@@ -29,22 +29,32 @@ Let's begin with the simplest example: a math agent with a tool call. This is a
2929
```python
3030
ajet --conf ./tutorial/example_math_agent/math_agent.yaml --backbone='verl'
3131
```
32+
<div align="center">
33+
<img width="640" alt="image" src="https://serve.gptacademic.cn/publish/shared/Image/classic+swarm+revise.jpg"/>
34+
</div>
3235

3336
### Swarm Mode
3437

38+
Let's begin with the simplest AgentJet Swarm example: also a math agent. In this case, you can use any GPU-less laptop to train the model remotely.
39+
3540
1. Start swarm server and begin swarm overwatch: `ajet-swarm start` and `ajet-swarm overwatch --swarm-url=http://localhost:10086`.
36-
2. From another device (or localhost), run [this script to train](https://github.com/modelscope/AgentJet/blob/main/tutorial/example_math_swarm/math.py):
41+
2. From your laptop (or swarm server localhost), run [this script to train](https://github.com/modelscope/AgentJet/blob/main/tutorial/example_math_swarm/math.py) to start training:
3742
```python
3843
AJET_SWARM_URL="http://swarm-server-ip:10086" python ./tutorial/example_math_swarm/math.py
3944
```
45+
<div align="center">
46+
<img width="600" alt="image" src="https://serve.gptacademic.cn/publish/shared/Image/swarm-server.gif"/>
47+
</div>
48+
4049

4150
## ✈️ Features
4251

4352
We aim to build a easy-to-learn Agent tuner that unlock more possibilities for agent developers:
4453

4554
- **Easy and Friendly**. AgentJet helps you tune models behind your agent workflows easily, optimizing your agents for top performance with minimal effort.
4655
- **Rich Tutorial Library**. AgentJet provides a rich library of [examples](https://github.com/modelscope/AgentJet/tree/main/tutorial) as tutorials.
47-
- **Efficient and Scalable**. AgentJet uses [verl] as the default backbone (`--backbone=verl`). However, we also support [trinity](https://github.com/modelscope/Trinity-RFT/) as alternative backbone, accelerating your tuning process via fully asynchronous RFT.
56+
- **Swarm Training**. [This unique feature](https://modelscope.github.io/AgentJet/en/swarm_intro_blog_english/) of AgentJet opens many possibilities: deploying distributed & self-healing rollout workers, **non-shared-parameter multi-agent** training, **multi-runtime & multi-task cocktail** training. And just like Tinker, you can use AgentJet Swarm to train **models even on **GPU-less laptop(s)**.
57+
- **Efficient and Scalable**. AgentJet uses [verl] as the default backbone (`--backbone=verl`). However, we also support trinity as alternative backbone, accelerating your tuning process via fully asynchronous RFT.
4858
- **Flexible and Fast**. AgentJet supports [multi-agent workflows](https://modelscope.github.io/AgentJet/en/workflow/) and adopts a context merging technique, accelerating training by 1.5x to 10x when the workflow involves multi-turn (or multi-agent) conversations.
4959
- **Reliability and Reproducibility**. Our team keeps track of framework performance across multiple [tasks + major-git-version + training-backbones](https://benchmark.agentjet.top/) (under construction, still gathering data, coming soon).
5060

@@ -62,13 +72,6 @@ For advanced researchers, AgentJet also provides high-resolution logging and deb
6272

6373
- **Click here to read the** [**installation guide**](https://modelscope.github.io/AgentJet/en/installation/).
6474

65-
#### Run Training
66-
67-
- You can start training your first agent with a single command using a pre-configured YAML file. Take the [Math agent](https://modelscope.github.io/AgentJet/en/example_math_agent/) as an example:
68-
69-
```bash
70-
ajet --conf tutorial/example_math_agent/math_agent.yaml
71-
```
7275

7376
#### Example Library
7477

@@ -111,6 +114,7 @@ The internal system orchestrates several specialized modules to handle the compl
111114
* **Task Runner**: Executes the Agent workflow and calculates rewards.
112115
* **Model Tuner**: Forwards inference requests from the workflow to the LLM engine.
113116
* **Context Tracker**: Monitors LLM calls and automatically merges shared-history timelines to improve training efficiency by **1.5x to 10x**.
117+
* **Swarm Server**: A data interchange center that accept OpenAI-like requests and engine instructions, activated only in AgentJet Swarm mode.
114118

115119

116120

@@ -128,14 +132,11 @@ AgentJet is a constantly evolving project. We are planning to add the following
128132

129133
| Category | Feature | Status |
130134
| :--- | :--- | :--- |
131-
| **Examples** | Covering LangGraph and AutoGen frameworks | Done & Verifying |
132135
| **Examples** | Add LoRA training examples | Todo |
133-
| **Infra** | Cross-process Tuner wrapper to pass though process forking | Done & Verifying |
134136
| **Infra** | Optimize configurations for long-context adaptation on smaller GPUs | In Progress |
135-
| **Capability** | Prompt tuning | In Progress |
136137
| **Capability** | Multi-modal training support | Todo |
137138
| **Capability** | MARL Credit assignment | Todo |
138-
| **Capability** | Training dataset generation from few-shot samples | Done & Verifying |
139+
| **Capability** | Training dataset generation from few-shot samples | Todo |
139140

140141

141142
## ✈️ Citation
@@ -158,8 +159,9 @@ If you use AgentJet in your research, please cite:
158159

159160
---
160161
<div align="center">
162+
This project is under active development, we need your help to make it shine! <br/>
161163

162-
[⭐ Star Us](https://github.com/modelscope/AgentJet) · [Report Bug](https://github.com/modelscope/AgentJet/issues) · [Request Feature](https://github.com/modelscope/AgentJet/issues)
164+
[⭐ Star Us](https://github.com/modelscope/AgentJet) · [✈️ Report Bug](https://github.com/modelscope/AgentJet/issues) · [✈️ Request Feature](https://github.com/modelscope/AgentJet/issues)
163165
</div>
164166

165167

File renamed without changes.
Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@ Unlike traditional setups where the user code must run inside the training clust
3131
| **架构模式** | 托管服务 + 单点客户端 API | **✅ 服务器和客户端都可按需拓展** |
3232
| **多客户端共同参与训练** | ❌ 不支持 | **✅ 支持** |
3333
| **远程算力部署方式** | Thinking Machines Lab 公司提供定价 | **✅ 自建 GPU 服务器端 或 使用阿里云灵骏** |
34-
| **训练方式** | LoRA 微调 | **✅ 全量 LLM 模型训练** |
34+
| **训练方式** | 仅限 LoRA 微调 | **✅ 全量 LLM 模型训练** |
3535
| **支持的模型** | ❌ 少部分 LLM 模型 | **✅ 大多数新旧 LLM 模型** |
3636
| **最大模型规模** | Llama 70B、Qwen 235B | **✅ 取决于用户 GPU 集群配置** |
3737
| **通信协议** | 专有 API | **✅ 专有API + OpenAI兼容API** |

0 commit comments

Comments
 (0)