|
| 1 | +from ajet.copilot.job import AgentJetJob |
| 2 | +from ajet.tuner_lib.experimental.as_swarm_client import SwarmClient, run_episodes_until_all_complete |
| 3 | +from ajet.default_config.ajet_default import AjetTaskReader |
| 4 | +from ajet.task_reader import RouterTaskReader |
| 5 | +from .frozenlake import FrozenLake |
| 6 | + |
| 7 | +import asyncio |
| 8 | +import threading |
| 9 | + |
| 10 | +# step 1: ajet-swarm start --swarm-port=10086 |
| 11 | +# step 2: ajet-swarm start --swarm-port=10087 |
| 12 | +# step 3: python -m tutorial.example_frozenlake_swarm.frozen_lake_roll |
| 13 | + |
| 14 | +# --------- configurations that take effect locally ------------- |
| 15 | +LOCAL_GRPO_N = 4 # grpo group size |
| 16 | +LOCAL_NUM_EPOCH = 10000 |
| 17 | + |
| 18 | +# --------- configurations that take effect remotely ------------- |
| 19 | +REMOTE_BATCH_SIZE = 32 |
| 20 | +REMOTE_1_SWARM_URL = "http://localhost:10086" # Change to your swarm remote url |
| 21 | +REMOTE_1_ALLOCATE_GPU_PER_NODE = 4 |
| 22 | +REMOTE_1_TRAIN_MODEL = '/mnt/data_cpfs/model_cache/modelscope/hub/Qwen/Qwen/Qwen2.5-7B-Instruct' |
| 23 | +REMOTE_2_SWARM_URL = "http://localhost:10087" # Change to your swarm remote url |
| 24 | +REMOTE_2_ALLOCATE_GPU_PER_NODE = 4 |
| 25 | +REMOTE_2_TRAIN_MODEL = '/mnt/data_cpfs/model_cache/modelscope/hub/Qwen/Qwen/Qwen2.5-3B-Instruct' |
| 26 | + |
| 27 | +class WeightUpdatedHalfway(Exception): |
| 28 | + """Raised when the remote side starts updating model weights halfway through an episode.""" |
| 29 | + |
| 30 | + |
| 31 | +def main(): |
| 32 | + |
| 33 | + dataset = RouterTaskReader(reader_type = "random_dummy", reader_config = AjetTaskReader()) |
| 34 | + |
| 35 | + # Hand shake with remote swarm server |
| 36 | + swarm_worker_7B = SwarmClient(REMOTE_1_SWARM_URL) |
| 37 | + swarm_worker_7B.auto_sync_train_config_and_start_engine( |
| 38 | + AgentJetJob( |
| 39 | + algorithm="grpo", |
| 40 | + project_name="ajet-swarm", |
| 41 | + experiment_name="test", |
| 42 | + n_gpu=REMOTE_1_ALLOCATE_GPU_PER_NODE, |
| 43 | + model=REMOTE_1_TRAIN_MODEL, |
| 44 | + batch_size=REMOTE_BATCH_SIZE, |
| 45 | + num_repeat=LOCAL_GRPO_N, |
| 46 | + ), |
| 47 | + ) |
| 48 | + # Hand shake with remote swarm server |
| 49 | + swarm_worker_3B = SwarmClient(REMOTE_2_SWARM_URL) |
| 50 | + swarm_worker_3B.auto_sync_train_config_and_start_engine( |
| 51 | + AgentJetJob( |
| 52 | + algorithm="grpo", |
| 53 | + project_name="ajet-swarm", |
| 54 | + experiment_name="test2", |
| 55 | + n_gpu=REMOTE_2_ALLOCATE_GPU_PER_NODE, |
| 56 | + model=REMOTE_2_TRAIN_MODEL, |
| 57 | + batch_size=REMOTE_BATCH_SIZE, |
| 58 | + num_repeat=LOCAL_GRPO_N, |
| 59 | + ), |
| 60 | + ) |
| 61 | + def play_different_swarm_server(task, swarm_worker:SwarmClient) -> float | None: |
| 62 | + # begin episode |
| 63 | + episode_uuid, api_baseurl_key = swarm_worker.begin_episode(discard_episode_timeout=120) |
| 64 | + # execute agent ( base_url = api_baseurl_key.base_url, api_key = api_baseurl_key.api_key ) |
| 65 | + env = FrozenLake( |
| 66 | + env_max_steps=20, |
| 67 | + agent_max_steps=20, |
| 68 | + seed=task.metadata["random_number"], |
| 69 | + ) |
| 70 | + workflow_output = asyncio.run(env.execute(task, api_baseurl_key.api_key, api_baseurl_key.base_url)) |
| 71 | + # report output back to swarm remote |
| 72 | + swarm_worker.end_episode(task, episode_uuid, workflow_output) |
| 73 | + # print global rollout status across the swarm |
| 74 | + swarm_worker.print_rollout_stat() |
| 75 | + return workflow_output.reward |
| 76 | + |
| 77 | + def rollout(task): |
| 78 | + f1 = threading.Thread(target=play_different_swarm_server, args=(task, swarm_worker_7B), daemon=True) |
| 79 | + f1.start() |
| 80 | + f2 = threading.Thread(target=play_different_swarm_server, args=(task, swarm_worker_3B), daemon=True) |
| 81 | + f2.start() |
| 82 | + f1.join() |
| 83 | + f2.join() |
| 84 | + return |
| 85 | + |
| 86 | + |
| 87 | + next_batch = [] |
| 88 | + for epoch in range(LOCAL_NUM_EPOCH): |
| 89 | + for _, task in enumerate(dataset.generate_training_tasks()): |
| 90 | + for _ in range(LOCAL_GRPO_N): |
| 91 | + next_batch.append(task) |
| 92 | + if len(next_batch) >= (REMOTE_BATCH_SIZE * LOCAL_GRPO_N): |
| 93 | + # wait until getting `local_batching_size` next_batch, then execute them with with retry logic |
| 94 | + run_episodes_until_all_complete(next_batch, func=rollout, auto_retry=True) |
| 95 | + next_batch.clear() |
| 96 | + return None |
| 97 | + |
| 98 | + |
| 99 | +if __name__ == "__main__": |
| 100 | + main() |
0 commit comments