Skip to content

Binex

Debuggable runtime for AI agent pipelines

Binex orchestrates multi-agent workflows defined in YAML. It executes DAG-based pipelines with any combination of LLM providers, records every step, and lets you trace, replay, debug, and diff runs — giving you full observability into your AI agent systems.

Key Features

  • YAML-defined workflows — Describe multi-agent pipelines as directed acyclic graphs with a simple, declarative format. No code required.
  • Multi-provider LLM support — Mix OpenAI, Anthropic, Gemini, Ollama, Groq, Mistral, DeepSeek, Together, and OpenRouter in a single workflow via LiteLLM routing.
  • Full run observability — Every node execution is recorded. Trace timelines, inspect artifacts, debug failures, and replay past runs.
  • Root-cause analysis — Automatic failure diagnosis with cascade detection, latency anomaly flagging, and actionable recommendations.
  • Run bisection — Find the exact divergence point between two runs with content similarity analysis.
  • Run diffing — Compare two workflow runs side-by-side to understand what changed between executions.
  • Output schema validation — Define JSON Schema for node outputs with automatic retry on validation failure.
  • Streaming LLM output — Watch LLM tokens arrive in real-time with auto-detection for TTY terminals.
  • Agent-to-Agent (A2A) protocol — Connect to remote A2A-compatible agent servers with built-in Gateway proxy for capability-based routing, automatic failover, and health monitoring.
  • Framework adapters — Integrate LangChain chains, CrewAI crews, and AutoGen teams as workflow nodes via thin-wrapper adapters. Install only what you need with optional extras.
  • Plugin system — Extend Binex with custom adapters using Python entry points or inline adapter_class configuration.
  • OpenTelemetry tracing — Optional run-level and node-level spans for external collectors (Jaeger, Tempo), with zero overhead when disabled.
  • Workflow versioning — Schema versioning with migration framework, plus workflow snapshots stored in SQLite for run reproducibility.
  • Export & webhooks — Export run data to CSV/JSON, webhook notifications on run lifecycle events.
  • Interactive CLI — Project scaffolding, workflow validation, a built-in doctor command, and a start wizard to get you productive quickly.

Install

pip install -e ".[dev]"

Quick Demo

binex hello                    # run a built-in demo workflow
binex run examples/simple.yaml # run a sample pipeline
binex debug <run-id>           # inspect the completed run

See the Quickstart for a full walkthrough.

Documentation

Section Description
Quickstart Install Binex and run your first workflow in under 5 minutes
CLI Reference All commands: hello, init, run, debug, trace, replay, diff, artifacts, dev, doctor, validate, scaffold, cancel, start, explore, diagnose, bisect, gateway, plugins, export, workflow
Concepts Core concepts: agents, workflows, artifacts, execution model, lineage tracking
Architecture Runtime internals: orchestrator, stores, adapters, scheduler, DAG engine
Workflow Format YAML schema reference with node specs, variables, conditionals, and defaults
Multi-Provider LLM Using multiple LLM providers in a single workflow
Contributing Development setup, testing guide, and code style

License

Binex is released under the MIT License.