Skip to content

Commit ce379be

Browse files
abrichrclaude
andauthored
feat: add GPU instance lifecycle CLI for model serving (#189)
Add `openadapt-gpu` CLI that automates the full GPU experiment workflow: launch EC2 instance, install inference engine (SGLang/vLLM), serve model, set up SSH tunnel, and terminate. Replaces 9 manual steps with one command. Commands: launch, status, terminate, serve, run-comparison, ssh, logs. State persisted to ~/.openadapt/gpu_state.json so terminate always works even when SSH is broken (uses EC2 API directly). Co-authored-by: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
1 parent e787d0c commit ce379be

2 files changed

Lines changed: 471 additions & 0 deletions

File tree

pyproject.toml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -116,6 +116,8 @@ openadapt-train-grpo = "scripts.train_trl_grpo:main"
116116
openadapt-eval = "scripts.run_full_eval:main"
117117
openadapt-collect = "scripts.collect_distillation_data:main"
118118
openadapt-analyze = "openadapt_evals.analysis.cli:main"
119+
# GPU instance lifecycle (launch, serve, terminate)
120+
openadapt-gpu = "scripts.gpu_cli:main"
119121

120122
[project.urls]
121123
Homepage = "https://github.com/OpenAdaptAI/openadapt-evals"

0 commit comments

Comments
 (0)