@@ -143,19 +143,25 @@ cat output.txt | squeez "Fix the CSRF validation bug"
143143
144144# Or with a file
145145squeez " Fix the CSRF bug" --input-file output.txt
146+
147+ # Explicit extract subcommand also works
148+ squeez extract " Fix the CSRF bug" --input-file output.txt
146149```
147150
148151### Python API
149152
150153``` python
151154from squeez.inference.extractor import ToolOutputExtractor
152155
153- # Connects to vLLM server (default: localhost:8000)
156+ # Load model from config/env
154157extractor = ToolOutputExtractor()
155158
156159# Or load model locally
157160extractor = ToolOutputExtractor(model_path = " ./output/squeez_qwen" )
158161
162+ # Or connect to a server explicitly
163+ extractor = ToolOutputExtractor(base_url = " http://localhost:8000/v1" , model_name = " squeez" )
164+
159165filtered = extractor.extract(
160166 task = " Fix the CSRF validation bug in middleware" ,
161167 tool_output = raw_output,
@@ -171,17 +177,27 @@ Backend is resolved in order: CLI args > env vars > config file (`squeez.yaml` o
171177
172178``` yaml
173179# squeez.yaml
174- model_path : " ./output/squeez_qwen" # local transformers
175- # base_url: "https://api.groq.com/openai/v1" # or remote API
180+ backend : " transformers" # optional preference
181+ local_model_path : " ./output/squeez_qwen"
182+ # server_url: "https://api.groq.com/openai/v1"
183+ # server_model: "squeez"
176184```
177185
178186``` bash
179187# Or via environment variables
180- export SQUEEZ_MODEL_PATH=./output/squeez_qwen
181- export SQUEEZ_BASE_URL=https://api.groq.com/openai/v1
188+ export SQUEEZ_LOCAL_MODEL=./output/squeez_qwen
189+ export SQUEEZ_SERVER_URL=https://api.groq.com/openai/v1
190+ export SQUEEZ_SERVER_MODEL=squeez
182191export SQUEEZ_API_KEY=gsk_...
183192```
184193
194+ Clear flag names are available on the CLI, with the old names kept as aliases:
195+
196+ ``` bash
197+ squeez " Fix the bug" --local-model ./output/squeez_qwen
198+ squeez " Fix the bug" --server-url http://localhost:8000/v1 --server-model squeez
199+ ```
200+
185201### Use with Claude Code
186202
187203Add this to your project's ` CLAUDE.md ` (or ` ~/.claude/CLAUDE.md ` for global):
@@ -216,7 +232,7 @@ This pulls the [SWE-bench tool output dataset](https://huggingface.co/datasets/K
216232### 2. Train with LoRA
217233
218234``` bash
219- python -m squeez.training. train \
235+ squeez train \
220236 --train-file data/train.jsonl \
221237 --eval-file data/eval.jsonl
222238```
@@ -226,8 +242,8 @@ Default: Qwen 3.5 2B with LoRA (r=16, alpha=32). See `configs/default.yaml` for
226242### 3. Evaluate
227243
228244``` bash
229- python -m squeez.training.evaluate \
230- --model-path output/squeez_qwen \
245+ squeez eval \
246+ --extractor-model output/squeez_qwen \
231247 --eval-file data/eval.jsonl
232248```
233249
@@ -275,11 +291,11 @@ Built from 2,294 [SWE-bench](https://huggingface.co/datasets/princeton-nlp/SWE-b
275291To regenerate the dataset from scratch:
276292
277293``` bash
278- python -m squeez.data. pipeline --phase all \
294+ squeez pipeline --phase 1 2 3 4 5 6 7 8 \
279295 --output-dir data \
280296 --github-token $GITHUB_TOKEN \
281- --openai -api-key $GROQ_API_KEY \
282- --distillation -base-url https://api.groq.com/openai/v1
297+ --teacher -api-key $GROQ_API_KEY \
298+ --teacher -base-url https://api.groq.com/openai/v1
283299```
284300
285301## Citation
0 commit comments