Explore runnable examples that show how to use Weco to optimize ML models, prompts, and GPU kernels. Pick an example and get going in minutes.
- Install the CLI
pip install weco| Example | Focus | Dependencies | Docs |
|---|---|---|---|
| 🧭 Hello World | Learn the Weco workflow on a small PyTorch model | torch |
README • Colab |
| 📋 LangSmith ZephHR QA | LLM-judge prompt optimization on HR policy QA | openai, langsmith, OpenAI + LangSmith API keys |
README |
| ⚡ Triton Optimization | Speed up attention with Triton kernels | numpy, torch, triton, NVIDIA GPU |
README |
| 🚀 CUDA Optimization | Generate low-level CUDA kernels for max speed | ninja, numpy, torch, triton, NVIDIA GPU, CUDA Toolkit |
README |
| 🧠 Prompt Engineering | Iteratively refine LLM prompts to improve accuracy | openai, datasets, OpenAI API key |
README |
| 📊 Agentic Scaffolding | Optimize agentic scaffolding for chart-to-CSV extraction | openai, huggingface_hub, uv, OpenAI API key |
README |
| 🛰️ Spaceship Titanic | Improve a Kaggle model training pipeline | pandas, numpy, scikit-learn, torch, xgboost, lightgbm, catboost |
README |
Minimal commands to run each example. For full context and explanations, see the linked READMEs.
Tip: Add
--apply-changeto any command below to automatically apply the best solution to your source file without prompting.
cd examples/hello-world
pip install -r requirements.txt
weco run --source module.py \
--eval-command "python evaluate.py --path module.py" \
--metric speedup \
--goal maximize \
--steps 15 \
--additional-instructions "Fuse operations in the forward method while ensuring the max float deviation remains small. Maintain the same format of the code."- Tip: Use
--device cuda(NVIDIA GPU) or--device mps(Apple Silicon).
- Requirements: OpenAI + LangSmith API keys
- Setup: Configure
helpfulness(1–5) andcorrectness(binary) online evaluators in LangSmith dashboard - Run:
cd examples/langsmith-zephhr-qa
python setup_dataset.py
weco run --source agent.py \
--eval-backend langsmith \
--langsmith-dataset zephhr-qa-opt \
--langsmith-target agent:answer_hr_question \
--langsmith-evaluators evaluators:json_schema_validity evaluators:conciseness \
--langsmith-dashboard-evaluators helpfulness correctness \
--langsmith-metric-function evaluators:qa_score \
--additional-instructions optimizer_exemplars.md \
--metric qa_score --goal maximize --steps 30- Requirements: NVIDIA GPU
cd examples/triton
pip install -r requirements.txt
weco run --source module.py \
--eval-command "python evaluate.py --path module.py" \
--metric speedup \
--goal maximize \
--steps 15 \
--model o4-mini \
--additional-instructions "Use a combination of triton and pytorch to optimize the forward pass while ensuring a small max float diff. Maintain the same code interface. Do not use any fallbacks. Assume any required dependencies are installed and data is already on the gpu." \
--eval-timeout 120- Requirements: NVIDIA GPU and CUDA Toolkit
- Optional: If compatible, install flash attention (
pip install flash-attn --no-build-isolation)
cd examples/cuda
pip install -r requirements.txt
weco run --source module.py \
--eval-command "python evaluate.py --path module.py" \
--metric speedup \
--goal maximize \
--steps 50 \
--model gpt-5 \
--additional-instructions "Write in-line CUDA using pytorch's load_inline() to optimize the code while ensuring a small max float diff. Maintain the same code interface. Do not use any fallbacks and never use the build_directory arg for load_inline(). Assume any required dependencies are installed and data is already on the gpu." \
--eval-timeout 600- Requirements: OpenAI API key (create here)
- Install Dependencies:
pip install openai datasets - Run:
cd examples/prompt
export OPENAI_API_KEY="your_key_here"
weco run --source optimize.py \
--eval-command "python eval.py" \
--metric score \
--goal maximize \
--steps 20 \
--model o4-mini \
--additional-instructions "Improve the prompt to get better scores. Focus on clarity, specificity, and effective prompt engineering techniques."- Requirements: OpenAI API key (create here)
- Install Dependencies:
pip install uv openai huggingface_hub - Run:
cd examples/extract-line-plot
export OPENAI_API_KEY="your_key_here"
uv run --with huggingface_hub python prepare_data.py # prepare dataset
weco run --source optimize.py \
--eval-command 'uv run --with openai python eval.py --max-samples 100 --num-workers 50' \
--metric accuracy \
--goal maximize \
--steps 20 \
--model gpt-5- Install Dependencies:
pip install pandas numpy scikit-learn torch xgboost lightgbm catboost - Run:
cd examples/spaceship-titanic
weco run --source train.py \
--eval-command "python evaluate.py --data-dir ./data --seed 0" \
--metric accuracy \
--goal maximize \
--steps 10 \
--model o4-mini \
--additional-instructions "Improve feature engineering, model choice and hyper-parameters." \
--log-dir .runs/spaceship-titanicIf you're new to Weco, start with Hello World, then try LangSmith ZephHR QA for a realistic LangSmith optimization workflow, explore Triton and CUDA for kernel engineering, Prompt Engineering for optimzing an LLM's prompt, Extract Line Plot for optimzing agentic scaffolds, or Spaceship Titanic for model development.