Skip to content

Latest commit

 

History

History
261 lines (178 loc) · 9.18 KB

File metadata and controls

261 lines (178 loc) · 9.18 KB

Claw Runtime Comparison

This note compares the eight runtime families supported by the Claw bridge layer:

  • OpenClaw
  • Hermes
  • OpenFang
  • NanoBot
  • PicoClaw
  • NanoClaw
  • ZeroClaw
  • NullClaw

The comparison is based on the current claw/ bridge design and runtime metadata. It is a practical analysis, not a formal benchmark report.

What The Bridge Optimizes For

The bridge is trying to satisfy eight different operating modes with one shared BitShares surface:

  • very small footprint and fast startup
  • broad assistant and plugin coverage
  • CLI-first workspace integration
  • simple MCP-based integration
  • low-cost hardware and launcher-based workflows
  • workspace-native skill loading and local manifest workflows

That means each runtime is a different compromise, not a strict upgrade path.

Comparison Axes

The useful axes are:

  • runtime footprint
  • startup latency
  • integration style
  • developer ergonomics
  • ecosystem breadth
  • operational complexity

Deep Dive

OpenClaw

OpenClaw is the broadest and heaviest option.

Strengths:

  • Richest runtime surface.
  • Native plugin model is a strong fit for extension-heavy workflows.
  • Broadest assistant-style experience among the seven.
  • Best when the bridge needs to live inside a feature-rich product rather than a small runtime shim.

Tradeoffs:

  • Highest operational complexity.
  • Largest dependency and runtime footprint.
  • More moving parts means more integration surface to keep aligned.

Best fit:

  • Desktop or server environments where footprint is less important.
  • Teams that value extensibility and a mature assistant ecosystem.
  • Workflows that benefit from plugin registration over a narrow CLI surface.

Hermes

Hermes is the best fit when Claw should be one capability inside a broader assistant runtime rather than the whole product.

Strengths:

  • Broad agent platform with memory, messaging, cron, browser automation, voice, and subagent delegation.
  • MCP-first integration lets Hermes consume the shared Claw surface without vendoring DEXBot2-specific runtime code.
  • Good fit when trading is only one part of a larger assistant workflow.

Tradeoffs:

  • Heavier operational surface than the narrower runtimes.
  • Adds little value if the only requirement is safe DEXBot inspection and execution.
  • The integration should stay MCP-first; a Hermes-specific bridge wrapper would mostly duplicate existing Claw behavior.

Best fit:

  • Users who want a general-purpose assistant that can also trade.
  • Always-on or messaging-driven workflows where memory and scheduling matter.
  • Teams that want Hermes features without moving BitShares execution outside DEXBot2.

OpenFang

OpenFang is the CLI-first local workspace integration with the thinnest maintenance surface.

Strengths:

  • Local CLI bridge keeps the integration narrow and explicit.
  • Generated SKILL.md can wrap the shared Claw surface without vendoring OpenFang internals.
  • Good fit when the runtime should consume commands from a workspace skill file instead of embedding Claw as a library.

Tradeoffs:

  • Less feature-rich than a full plugin-driven assistant platform.
  • Depends on OpenFang's local CLI and workspace skill conventions.
  • Best when the bridge stays deliberately thin.

Best fit:

  • Local workspace installs.
  • Operators who want the smallest possible adapter between OpenFang and DEXBot2.
  • CLI-first workflows that do not need a new transport layer.

NanoBot

NanoBot sits between OpenClaw and PicoClaw in spirit, but it is a different tradeoff: a smaller Python codebase with MCP integration.

Strengths:

  • Easier to inspect and adapt than a large feature-rich assistant stack.
  • MCP gives a simple external-tool boundary.
  • Good for lightweight assistant workflows and rapid iteration.

Tradeoffs:

  • Slower startup and heavier runtime cost than Go or Rust.
  • Python dependency management adds operational friction compared with a static or near-static binary workflow.
  • Less suited to very small hardware or always-on resource-constrained deployments.

Best fit:

  • Teams that want a lightweight assistant they can reason about quickly.
  • Environments where Python ergonomics matter more than the absolute smallest footprint.
  • MCP-first integrations with moderate complexity.

PicoClaw

PicoClaw is the smallest Go-based option in the set and is the most launcher-oriented.

Strengths:

  • Small binary deployment model.
  • Good fit for low-cost boards and constrained Linux targets.
  • MCP-based tool integration keeps the external surface consistent.
  • Web launcher support makes it friendlier for desktop-style setup and usage.

Tradeoffs:

  • The platform is still evolving quickly, so behavior and footprint can shift more often.
  • Less mature than the longer-running ecosystems around OpenClaw or the broader upstream Rust tooling around ZeroClaw.
  • The launcher and workspace behavior add some setup expectations that are easy to miss without good docs.

Best fit:

  • Small-board deployments.
  • Users who want a low-footprint Go runtime with a practical launcher.
  • Deployments where the assistant needs to be easy to bootstrap, not just tiny.

NanoClaw

NanoClaw is the Claude Code skill-driven local runtime with a narrow bridge wrapper.

Strengths:

  • Uses a plain SKILL.md skill file and a local bridge script.
  • Fits the same shared Claw command surface without introducing a separate transport stack.
  • Keeps the bridge skill name distinct from NanoClaw's bundled claw skill.

Tradeoffs:

  • More workspace-convention dependent than the external-tool runtimes.
  • Newer than the longer-established OpenClaw, NanoBot, and PicoClaw flows.
  • Best experience depends on NanoClaw's .claude/skills layout and local CLI invocation.

Best fit:

  • Users who want a Claude Code style local assistant with a narrow bridge.
  • Workspace-driven installs where a generated bitshares-claw skill file is acceptable.
  • Operators who want to keep runtime integration simple and file-based.

ZeroClaw

ZeroClaw is the smallest and most constrained option.

Strengths:

  • Best cold-start behavior.
  • Best fit for small, static, local deployments.
  • Clear skill-manifest model with SKILL.toml.
  • Best when the bridge should feel tiny and deterministic rather than broad.

Tradeoffs:

  • Rust-oriented integration is more specialized than the other options.
  • The workflow is more manifest-driven, so the runtime surface is less flexible than plugin-first systems.
  • Better for fast, narrow automation than for a broad assistant shell.

Best fit:

  • Edge devices.
  • Minimal local automation.
  • Users who want the most predictable, lowest-overhead runtime.

NullClaw

NullClaw is the Zig-native runtime focused on workspace skill loading.

Strengths:

  • Native SKILL.toml loading in the workspace skill tree.
  • MCP server support alongside the built-in skill loader.
  • Strong fit for local, file-based assistant workflows.
  • Keeps the bridge surface aligned with the NullClaw workspace conventions.

Tradeoffs:

  • Newer integration path than the longer-established OpenClaw, NanoBot, and PicoClaw flows.
  • More workspace-centric than the CLI-first ZeroClaw wrapper.
  • Best experience depends on ~/.nullclaw/workspace/skills and NullClaw's config conventions.

Best fit:

  • Users who want a Zig-native assistant runtime with native skill manifests.
  • Local workflows where the bridge should slot into a NullClaw workspace directly.
  • Operators who want workspace-native NullClaw support without changing the shared Claw bridge surface.

Recommendation Matrix

If you optimize primarily for:

  • Lowest footprint: ZeroClaw
  • Broadest feature set: OpenClaw
  • General-purpose assistant that can also trade: Hermes
  • CLI-first workspace integration: OpenFang
  • Simplest external tool boundary: NanoBot or PicoClaw, depending on whether you want Python or Go
  • Skill-file driven local assistant: NanoClaw
  • Lowest-cost hardware: PicoClaw, then ZeroClaw
  • Most mature assistant ecosystem: OpenClaw
  • Fastest local command-style integration: ZeroClaw
  • Workspace-native skill manifests: NullClaw

Practical Rule Of Thumb

  • Choose OpenClaw if the bridge should live inside a larger assistant product with rich extension points.
  • Choose Hermes if you want a broader assistant runtime and want Claw to remain an MCP-backed trading capability inside it.
  • Choose OpenFang if you want a CLI-first workspace bridge with minimal maintenance overhead.
  • Choose NanoBot if you want a compact Python assistant that is easy to modify.
  • Choose PicoClaw if you want a small Go runtime with MCP and launcher support.
  • Choose NanoClaw if you want a Claude Code skill-driven local runtime with a narrow bridge.
  • Choose ZeroClaw if the bridge must be tiny and deterministic.
  • Choose NullClaw if you want a Zig-native runtime with workspace skill loading and MCP support.

Source Of Truth

The executable behavior lives in the claw/ modules and scripts. This comparison is documentation only and should be kept aligned with:

  • modules/claw_runtime_matrix.js
  • modules/claw_catalog.js
  • modules/claw_manifest.js
  • modules/claw_skill_md.js
  • modules/openfang_bridge.js
  • modules/nanoclaw_bridge.js
  • scripts/openfang_bridge.js
  • scripts/claw_skill_md.js
  • scripts/nanoclaw_bridge.js
  • scripts/zeroclaw_skill.js