Skip to content

Commit c23c096

Browse files
committed
Better flow
1 parent d622a3a commit c23c096

1 file changed

Lines changed: 9 additions & 22 deletions

File tree

README.md

Lines changed: 9 additions & 22 deletions
Original file line numberDiff line numberDiff line change
@@ -5,18 +5,22 @@
55
<br><em>Squeeze out the juice, leave the pulp behind.</em>
66
</p>
77

8-
LLM coding agents waste 80-95% of context tokens on irrelevant tool output. Squeez extracts only the lines that matter, compressing tool output by ~91% while keeping 86% of the relevant information.
9-
108
[![PyPI](https://img.shields.io/pypi/v/squeez)](https://pypi.org/project/squeez/)
119
[![Model](https://img.shields.io/badge/HF-Squeez--2B-yellow.svg)](https://huggingface.co/KRLabsOrg/squeez-2b)
1210
[![Dataset](https://img.shields.io/badge/HF-Dataset-yellow.svg)](https://huggingface.co/datasets/KRLabsOrg/tool-output-extraction-swebench)
1311
[![License](https://img.shields.io/badge/License-Apache%202.0-blue.svg)](https://opensource.org/licenses/Apache-2.0)
1412

15-
## How it works
13+
- Tool output pruner for LLM coding agents
14+
- Pipe any tool output (pytest, grep, git log, npm build, kubectl, ...) through squeez with a task description, get back only the relevant lines
15+
- Fine-tuned Qwen 3.5 2B, 0.79 F1, ~91% compression
16+
- CLI pipe, Python library, or vLLM server
1617

17-
Squeez uses a fine-tuned Qwen 3.5 2B model to read tool output alongside a task description and return only the relevant lines.
18+
```bash
19+
pip install squeez
20+
python -m pytest tests/ -v 2>&1 | squeez "find the test failure related to authentication"
21+
```
1822

19-
### Example: filtering test output
23+
## Example
2024

2125
Task: *"Find the test failure related to authentication"*
2226

@@ -81,10 +85,6 @@ E Got: rejection after 15m (timeout changed?)
8185
</tr>
8286
</table>
8387

84-
```bash
85-
$ python -m pytest tests/ -v 2>&1 | squeez "find the test failure related to authentication"
86-
```
87-
8888
<details>
8989
<summary><b>More examples</b></summary>
9090

@@ -134,28 +134,18 @@ Evaluated on 617 held-out test samples from SWE-bench, across 14 tool types:
134134

135135
Squeez-2B (2B params) outperforms a 35B MoE model at zero-shot and is 6x better than BM25 on Span F1.
136136

137-
## Install
138-
139-
```bash
140-
pip install squeez
141-
```
142-
143137
## Quick start
144138

145139
### With vLLM (recommended)
146140

147141
```bash
148-
# Start the server
149142
pip install vllm
150143
vllm serve KRLabsOrg/squeez-2b --dtype bfloat16 --max-model-len 16384
151144

152145
# Use from squeez CLI
153146
pip install squeez
154147
export SQUEEZ_SERVER_URL=http://localhost:8000/v1
155148
cat output.txt | squeez "find the bug"
156-
157-
# Or pipe directly
158-
python -m pytest tests/ -v 2>&1 | squeez "find the test failure related to authentication"
159149
```
160150

161151
vLLM keeps the model warm in memory with batched inference and high throughput.
@@ -192,9 +182,6 @@ extractor = ToolOutputExtractor()
192182
# Or connect to a server
193183
extractor = ToolOutputExtractor(base_url="http://localhost:8000/v1")
194184

195-
# Or use a custom local model
196-
extractor = ToolOutputExtractor(model_path="./output/squeez_qwen")
197-
198185
filtered = extractor.extract(
199186
task="Find the referer validation block",
200187
tool_output=raw_output,

0 commit comments

Comments
 (0)