Skip to content

Commit 8af0644

Browse files
committed
add logo
1 parent 7c70243 commit 8af0644

7 files changed

Lines changed: 54 additions & 143 deletions

File tree

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@ Let's begin with the simplest example: a math agent with a tool call.
2323
- First, please check out the [installation guide](https://doc.agentjet.top/AgentJet/en/installation/) to set up the training environment.
2424
- Then, tune your first model using the minimum example.
2525
```python
26-
ajet --conf tutorial/example_math_agent/math_agent.yaml --backbone='verl' --with-ray
26+
ajet --conf tutorial/example_math_agent/math_agent.yaml --backbone='verl'
2727
```
2828

2929

docs/en/configuration.md

Lines changed: 2 additions & 107 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22

33
This page provides a detailed description of the configuration files for AgentJet.
44

5-
---
5+
66

77
## Overview
88

@@ -31,7 +31,7 @@ At a high level, a typical config contains a single root section `ajet`, which i
3131

3232
</div>
3333

34-
---
34+
3535

3636
## Model Configuration
3737

@@ -58,7 +58,6 @@ export DASHSCOPE_API_KEY='sk-xxxxxx|sk-yyyyyy'
5858
export DASHSCOPE_API_KEY_BACKUP='sk-zzzzzz'
5959
```
6060

61-
---
6261

6362
## Data Configuration
6463

@@ -123,7 +122,6 @@ ajet:
123122
| `customized_protocol` | Use a custom Python class for scoring |
124123
| `rubrics_auto_grader` | Use LLM-based automatic grading |
125124

126-
---
127125

128126
## Training Configuration
129127

@@ -213,7 +211,6 @@ ajet:
213211
| `use_kl_loss` | Include KL divergence in loss |
214212
| `kl_loss_coef` | KL loss coefficient |
215213

216-
---
217214

218215
## Debug Mode
219216

@@ -234,7 +231,6 @@ ajet:
234231
- **Fixing randomness**: `debug_vllm_seed` helps reproduce issues
235232
- **Reduced parallelism**: Easier to debug with smaller concurrency
236233

237-
---
238234

239235
## Logging & Monitoring
240236

@@ -262,108 +258,7 @@ All experiment outputs are saved in `./launcher_record/{experiment_name}`:
262258
| **Metrics** | Training metrics (depends on logger) |
263259
| **Checkpoint** | Model checkpoints |
264260

265-
---
266-
267-
## Full Configuration Example
268-
269-
??? example "Complete Configuration Template"
270-
```yaml title="config.yaml"
271-
ajet:
272-
project_name: "ajet_default_project"
273-
experiment_name: "read_yaml_name"
274-
experiment_dir: "auto"
275-
backbone: debug
276-
277-
model:
278-
path: /path/to/model/Qwen2.5-14B-Instruct
279-
280-
data:
281-
max_prompt_length: 3000
282-
max_response_length: 15000
283-
train_batch_size: 32
284-
285-
rollout:
286-
user_workflow: tutorial.example_appworld.appworld->ExampleAgentScopeWorkflow
287-
force_disable_toolcalls: False
288-
max_env_worker: 128
289-
gamma: 1.0
290-
compute_madness_checklist:
291-
- "nonsense"
292-
agent_madness_termination: True
293-
agent_madness_reward: -1.0
294-
max_response_length_in_one_turn: 4096
295-
max_model_len: 18000
296-
multi_turn:
297-
max_sample_per_task: 30
298-
max_steps: 30
299-
expected_steps: 1
300-
tensor_model_parallel_size: 1
301-
n_vllm_engine: 2
302-
max_num_seqs: 10
303-
name: vllm
304-
num_repeat: 4
305-
temperature: 0.9
306-
top_p: 1.0
307-
val_kwargs:
308-
temperature: 0.0
309-
top_k: -1
310-
top_p: 1.0
311-
do_sample: False
312-
num_repeat: 1
313-
314-
task_reader:
315-
type: env_service
316-
env_service:
317-
env_type: "appworld"
318-
env_url: "http://127.0.0.1:8080"
319-
env_action_preference: code
320-
training_split: train
321-
validation_split: dev
322-
323-
task_judge:
324-
judge_type: customized_protocol
325-
judge_protocol: ajet.task_judge.env_service_as_judge->EnvServiceJudge
326-
alien_llm_model: qwen3-235b-a22b-instruct-2507
327-
alien_llm_response_length: 512
328-
329-
debug:
330-
debug_max_parallel: 16
331-
debug_first_n_tasks: 2
332-
debug_vllm_port: 18000
333-
debug_vllm_seed: 12345
334-
debug_tensor_parallel_size: 4
335-
336-
trainer_common:
337-
val_before_train: False
338-
val_pass_n: 4
339-
save_freq: 20
340-
test_freq: 20
341-
total_epochs: 50
342-
nnodes: 1
343-
n_gpus_per_node: 8
344-
logger: swanlab
345-
algorithm:
346-
adv_estimator: grpo
347-
use_kl_in_reward: False
348-
mini_batch_num: 1
349-
fsdp_config:
350-
param_offload: True
351-
optimizer_offload: True
352-
optim:
353-
lr: 1e-6
354-
use_kl_loss: True
355-
kl_loss_coef: 0.002
356-
kl_loss_type: low_var_kl
357-
ulysses_sequence_parallel_size: 1
358-
checkpoint_base_dir: ./saved_checkpoints
359-
360-
context_tracker:
361-
context_tracker_type: "linear"
362-
alien_llm_model: qwen3-235b-a22b-instruct-2507
363-
alien_llm_response_length: 512
364-
```
365261

366-
---
367262

368263
## Next Steps
369264

docs/en/installation.md

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -9,9 +9,10 @@ This document provides a step-by-step guide to installing AgentJet.
99

1010
## Prerequisites
1111

12-
| Requirement | Version |
12+
| Requirement | Detail |
1313
|-------------|---------|
14-
| **Python** | 3.10 |
14+
| **Python** | 3.10 |
15+
| Package Management | `uv` or `conda` |
1516

1617

1718
## Install from Source

docs/en/intro.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ AgentJet aims to build a state-of-the-art agent tuning platform for both develop
1212
- **Easy and Friendly**. AgentJet helps you tune models behind your agent workflows easily, optimizing your agents for top performance with minimal effort.
1313
- **Rich Tutorial Library**. AgentJet provides a rich library of [examples](https://github.com/modelscope/AgentJet/tree/main/tutorial) as tutorials.
1414
- **Efficient and Scalable**. AgentJet uses [verl] as the default backbone (`--backbone=verl`). However, we also support [trinity](https://github.com/modelscope/Trinity-RFT/) as alternative backbone, accelerating your tuning process via fully asynchronous RFT.
15-
- **Flexible and Fast**. AgentJet supports [multi-agent workflows](docs/en/workflow.md) and adopts a context merging technique, accelerating training by 1.5x to 10x when the workflow involves multi-turn (or multi-agent) conversations.
15+
- **Flexible and Fast**. AgentJet supports [multi-agent workflows](workflow.md) and adopts a context merging technique, accelerating training by 1.5x to 10x when the workflow involves multi-turn (or multi-agent) conversations.
1616
- **Reliability and Reproducibility**. Our team keeps track of framework performance across multiple [tasks + major-git-version + training-backbones](https://benchmark.agent-matrix.com/) (under construction, still gathering data, comming soon).
1717

1818
For advanced researchers, AgentJet also provides high-resolution logging and debugging solutions:

docs/en/quickstart.md

Lines changed: 18 additions & 29 deletions
Original file line numberDiff line numberDiff line change
@@ -1,39 +1,23 @@
11
# Quick Start
22

3-
AgentJet provides a complete feature set for tuning agents. You can try starting training an agent right away:
3+
4+
## 1. Testing Pre-define Demo
5+
6+
AgentJet provides a complete feature set for tuning agents. You can try starting training an agent right away by running a demo:
47

58
```bash
69
ajet --conf tutorial/example_math_agent/math_agent.yaml
710
```
811

9-
---
1012

11-
## Minimum Example
13+
## 2. Minimum Example
1214

1315
Let's begin with the simplest example: a math agent with a tool call.
1416

15-
<div class="workflow-single">
16-
<div class="workflow-header">Getting Started Flow</div>
17-
18-
<div class="workflow">
19-
<ol class="workflow-steps">
20-
<li><strong>Set up Environment</strong>
21-
22-
Check out the [installation guide](./installation.md) to set up the training environment.</li>
23-
<li><strong>Define Your Workflow</strong>
24-
25-
Write an Agent class (e.g., `MathToolWorkflow`) that inherits from the base Workflow class.</li>
26-
<li><strong>Configure and Run</strong>
27-
28-
Use the `AgentJetJob` API to configure and start training.</li>
29-
</ol>
30-
</div>
31-
</div>
32-
33-
### Code Example
34-
3517
```python title="train_math_agent.py"
3618
from ajet import AgentJetJob
19+
20+
# refer to `https://doc.agentjet.top/AgentJet/en/tune_your_first_agent/` on how to write your own workflow
3721
from tutorial.example_math_agent.math_agent_simplify import MathToolWorkflow
3822

3923
model_path = "YOUR_MODEL_PATH"
@@ -57,9 +41,17 @@ tuned_model = job.tune()
5741
ajet --conf ./saved_experiments/math.yaml
5842
```
5943

60-
---
6144

62-
## Explore Examples
45+
## 3. Compare with Community Training Curves
46+
47+
<div class="card-grid">
48+
<a href="https://benchmark.agent-matrix.com/examples" class="feature-card"><div class="card-header"><img src="https://api.iconify.design/mdi:calculator-variant.svg" class="card-icon card-icon-math" alt=""><h3>AgentJet Bechmark Tracking System (Developing-In-Progress)</h3></div><p class="card-desc">Compare training culves with community. Investigate the influence of versions, backbones, hyper-parameters, etc.</p></a>
49+
</div>
50+
51+
52+
53+
54+
## 4. Explore Example Gallery
6355

6456
Explore our rich library of examples to kickstart your journey:
6557

@@ -72,11 +64,8 @@ Explore our rich library of examples to kickstart your journey:
7264
<a href="./example_frozenlake/" class="feature-card"><div class="card-header"><img src="https://api.iconify.design/mdi:snowflake.svg" class="card-icon card-icon-data" alt=""><h3>Frozen Lake</h3></div><p class="card-desc">Solving a frozen lake walking puzzle.</p></a>
7365
</div>
7466

75-
---
76-
77-
---
7867

79-
## Next Steps
68+
## 5. Next Steps
8069

8170
<div class="card-grid">
8271
<a href="../tune_your_first_agent/" class="feature-card"><div class="card-header"><img src="https://api.iconify.design/mdi:rocket-launch.svg" class="card-icon card-icon-agent" alt=""><h3>Tune Your First Agent</h3></div><p class="card-desc">Complete step-by-step guide to building your own agent from scratch.</p></a>

docs/en/tune_your_first_agent.md

Lines changed: 28 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -51,7 +51,7 @@ tutorial/example_math_agent
5151
└── math_agent.yaml
5252
```
5353

54-
Next, define your workflow (or convert an existing workflow). Here we use AgentScope to implement this agent. You can toggle two code before and after convertion to see the difference. If you prefer langchain or openai sdk, [please refer to this article](../agent_framework_support).
54+
Next, define your workflow (or convert an existing workflow). Here we use AgentScope to implement this agent. You can toggle two code before and after convertion to see the difference. If you prefer langchain or openai sdk, [please refer to this article](agent_framework_support.md).
5555

5656
=== "`math_agent.py` - AgentJet Workflow (After Convertion)"
5757

@@ -199,6 +199,32 @@ Now, we have obtained all materials required to train the agent.
199199

200200
```
201201

202+
### Configuration Parameters
203+
204+
| Category | Parameter | Description | Example Value |
205+
|----------|-----------|-------------|---------------|
206+
| **Project** | `project_name` | Name of the training project | `example_math_agent` |
207+
| **Task Reader** | `type` | Type of data source to read tasks from | `huggingface_dat_repo` (options: `env_service`, `dataset_file`, `huggingface_dat_repo`) |
208+
| | `dataset_path` | Path or identifier of the dataset | `openai/gsm8k` |
209+
| | `training_split` | Dataset split used for training | `train` |
210+
| | `validation_split` | Dataset split used for validation/testing | `test` |
211+
| **Model** | `path` | Path or identifier of the model to be trained | `Qwen/Qwen2.5-7B` |
212+
| **Rollout** | `user_workflow` | Python module path to the workflow class | `tutorial.example_math_agent.math_agent->ExampleMathLearn` |
213+
| | `num_repeat` | Number of rollout repeats per task (GRPO `n` parameter) | `6` |
214+
| | `tensor_model_parallel_size` | vLLM tensor parallelism size | `1` |
215+
| | `max_response_length_in_one_turn` | Maximum token length for a single agent response | `1024` |
216+
| | `max_model_len` | Maximum total context length for the model | `10000` |
217+
| **Data** | `train_batch_size` | Number of tasks per training batch | `100` |
218+
| | `max_prompt_length` | Maximum token length for input prompts | `3000` |
219+
| | `max_response_length` | Maximum token length for model responses | `7000` |
220+
| **Debug** | `debug_max_parallel` | Maximum parallel workers in debug mode | `1` |
221+
| | `debug_first_n_tasks` | Number of tasks to process in debug mode | `1` |
222+
| **Trainer** | `save_freq` | Frequency (in steps) to save model checkpoints | `100` |
223+
| | `test_freq` | Frequency (in steps) to run validation | `100` |
224+
| | `total_epochs` | Total number of training epochs | `100` |
225+
| | `logger` | Logging backend for experiment tracking | `swanlab` |
226+
| **Task Judge** | `judge_protocol` | Protocol for judging task completion | `null` (reward is computed in workflow) |
227+
202228

203229
## Step 3: ✨Debug (Optional)
204230

@@ -231,7 +257,7 @@ We choose VSCode to debug because it is open-source and fast.
231257

232258
After `.vscode/launch.json` is created, press `F5` to start debugging. (Do not forget to configure python venv path in VSCode.)
233259

234-
For more debugging techniques, please refer to [debugging guidelines](../debugging_guide).
260+
For more debugging techniques, please refer to [debugging guidelines](debugging_guide.md).
235261

236262

237263
## Step 4: ✨Start Training

docs/index.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -82,7 +82,7 @@
8282
</div>
8383

8484

85-
We recommend using `uv` for dependency management. [Click here](./installation/) for details and other training backbone (e.g. Trinity-RFT) options.
85+
We recommend using `uv` for dependency management. [Click here](en/installation.md) for details and other training backbone (e.g. Trinity-RFT) options.
8686

8787
- Clone the Repository:
8888
```bash

0 commit comments

Comments
 (0)