WorkingAgents + Distributional: Discovering What Your Agents Are Actually Doing

By James Aspinwall, co-written by Alfred Pennyworth (my trusted AI) — March 7, 2026, 17:55


The Problem: You Don’t Know What You Don’t Know

Most AI observability tools answer questions you already thought to ask. “How many tool calls failed?” “What’s the average latency?” “Did the agent hallucinate?” These are important, but they assume you know where to look.

Distributional asks a different question: What behavioral patterns exist in your agent’s production data that you haven’t discovered yet?

WorkingAgents orchestrates 50+ MCP tools across CRM, task management, content, and communications. Every day, agents make thousands of decisions — which tool to call, what parameters to pass, how to synthesize results. Somewhere in that data are patterns that explain why some sessions succeed and others don’t, why certain users get better results than others, why performance drifts over time.

Distributional’s DBNL platform finds those patterns through unsupervised statistical analysis. It doesn’t require you to define what “good” looks like upfront. It discovers the behavioral fingerprint of your agents and surfaces deviations, clusters, and shifts you wouldn’t have thought to monitor.

This is a fundamentally different capability from what WorkingAgents has today — and from what most evaluation platforms offer.


What Distributional Brings

Distributional (DBNL) is an adaptive analytics platform for production AI agents. Founded in September 2023 by the SigOpt team (acquired by Intel in 2020), backed by $30M from Andreessen Horowitz, Two Sigma Ventures, and others, the platform is built on a core insight: AI behavior is probabilistic, not deterministic, and testing it requires statistical methods native to that reality.

The Distributional Fingerprint

Every AI application has what Distributional calls a “distributional fingerprint” — its unique baseline mixture of characteristic distributions across behavior dimensions. This fingerprint captures:

When the fingerprint shifts — a new topic cluster emerges, a tool sequence that used to work starts failing, latency correlates with a specific user segment — DBNL surfaces it as an Insight.

The Adaptive Analytics Flywheel

DBNL operates through an eight-step cycle:

IngestEnrichAnalyzePublishDiscoverInvestigateTrackRepeat

  1. Ingest: Production logs arrive via OpenTelemetry traces, SDK push, or SQL pull
  2. Enrich: Each log line is augmented with LLM-as-Judge evaluations, NLP metrics, topic classification, embeddings, and custom metrics — creating a rich behavioral vector per interaction
  3. Analyze: Unsupervised learning and statistical techniques discover patterns — temporal shifts, behavioral clusters, outliers
  4. Publish: Patterns appear as human-readable Insights and Dashboards
  5. Discover: Teams review automatically surfaced signals they didn’t know to look for
  6. Investigate: The Explorer tool enables population and temporal comparisons, drilling into the evidence behind each signal
  7. Track: Meaningful patterns become saved Segments and custom Metrics for ongoing monitoring
  8. Repeat: Tracked signals feed back into the enrichment and analysis, deepening future discovery

Three Types of Insights

Deployment Model

DBNL is free, open, and downloadable. It deploys in your Kubernetes cluster within your VPC. No data leaves your environment. Enterprise features include OpenID Connect SSO, role-based access, and workspace administration.

This matters. For an AI consulting firm deploying agents for clients, “your data stays in your infrastructure” eliminates the security objection before it’s raised.


What WorkingAgents Brings

WorkingAgents (“The Orchestrator”) is an Elixir OTP platform that gives AI agents real tools for business operations:

WorkingAgents generates exactly the kind of rich, multi-dimensional production data that Distributional is designed to analyze — tool calls, user interactions, topic diversity, model switching, and real business outcomes.


Where the Synergy Lives

1. Unsupervised Discovery on Tool Usage Patterns

WorkingAgents has 50+ tools. Users interact with agents in natural language, and the agent decides which tools to call, in what order, with what parameters. Today, there’s no systematic way to understand these tool usage patterns at scale.

Distributional’s unsupervised analysis would discover patterns like:

These are the patterns you wouldn’t think to monitor because you didn’t know they existed. Traditional observability counts tool calls. Distributional discovers the behavioral relationships between them.

2. User Behavior Segmentation

WorkingAgents serves different users with different needs. James manages CRM contacts. Jimmy asks about tasks and deadlines. Other consulting clients will have their own patterns. Distributional’s segment discovery would reveal:

This segmentation feeds directly into product decisions. If WhatsApp-originated tasks have higher completion rates, that’s a signal to invest more in the WhatsApp bridge experience.

3. Multi-Provider Model Comparison — Beyond Scores

Other evaluation platforms compare models with predefined metrics: accuracy, latency, cost. Distributional adds a dimension they can’t: behavioral fingerprint comparison.

WorkingAgents users can switch between Claude, OpenRouter models, and Perplexity at runtime. Distributional wouldn’t just score each provider — it would discover how the behavioral distribution changes:

This is richer than “Model A scored 4.2, Model B scored 3.8.” It reveals how models behave differently, not just how well.

4. The AI Data Flywheel for Consulting Clients

Distributional explicitly positions their platform around the “Analytics-Driven AI Data Flywheel” — using discovered signals and surfaced examples for post-training optimization. For WorkingAgents’ consulting business, this creates a concrete service offering:

Month 1: Deploy — Install WorkingAgents with custom tools for the client’s domain. Connect DBNL to ingest traces.

Month 2: Discover — DBNL surfaces behavioral patterns. “Your agents handle inventory queries well but struggle with multi-step procurement workflows. Here are the 47 example traces showing the failure pattern.”

Month 3: Optimize — Use the surfaced examples for prompt engineering, tool redesign, or model switching. DBNL’s tracked segments measure whether the changes worked.

Month 4+: Flywheel — Each optimization cycle surfaces new patterns in the changed behavior. The agent gets measurably better every month, with evidence.

This is recurring revenue built on data, not opinion. The consulting engagement doesn’t end after deployment — it becomes a continuous optimization service powered by Distributional’s discovery engine.

5. Permission and Access Pattern Analytics

WorkingAgents’ capability-based access control system creates a rich dataset: which users have which permissions, which tools they actually call, and how their usage patterns differ from their permission scope.

Distributional could surface insights like:

6. Temporal Drift Detection on Agent Behavior

AI agents aren’t static. Model updates, prompt changes, data drift, and user behavior evolution all cause the behavioral fingerprint to shift. WorkingAgents currently has no way to detect these shifts.

Distributional’s temporal insights would catch:

These temporal signals are invisible to snapshot-based evaluation tools. They only emerge from continuous distributional analysis.


Why Distributional is Different From Arize or Deepchecks

Distributional occupies a distinct position in the AI analytics landscape:

Dimension Arize AI Deepchecks Distributional
Core approach Trace observability Evaluation scoring Behavioral discovery
Primary question “What happened?” “Was it good?” “What patterns exist?”
Method OpenTelemetry spans LLM-as-Judge swarm Unsupervised statistical analysis
Requires predefined metrics Partially Yes No — discovers metrics
Deployment Cloud SaaS or self-hosted Cloud SaaS or on-prem Free, open, in your VPC
Best for Debugging specific failures Scoring agent quality Finding unknown unknowns

For WorkingAgents, these three platforms are complementary layers:

  1. Arize traces what happened (the execution path)
  2. Deepchecks evaluates whether it was done well (the quality score)
  3. Distributional discovers what you should be paying attention to (the behavioral signal)

Distributional fills the gap between “we monitor our agents” and “we understand our agents.”


The Gap Analysis

WorkingAgents Gap Distributional Solution
No behavioral pattern discovery Unsupervised learning surfaces unknown clusters, shifts, and outliers
No tool usage correlation analysis Distributional fingerprint captures tool-sequence-to-outcome correlations
No user segmentation analytics Segment Insights automatically cluster user behavior profiles
No temporal drift detection Temporal Insights surface behavioral shifts over time
No data flywheel for continuous improvement Adaptive Analytics Flywheel with surfaced examples for optimization
No multi-dimensional model comparison Behavioral fingerprint comparison across providers
Distributional Gap WorkingAgents Solution
Need production agent data sources 50+ MCP tool traces with rich business context
Need diverse tool-calling patterns CRM + tasks + content + communication tool chains
Need multi-provider comparison scenarios Runtime-switchable Claude/OpenRouter/Perplexity
Need consulting distribution channel AI consulting firm deploying for medium-size companies
Need non-Python ecosystem references Elixir OTP — unique agent orchestration stack
Need real business outcome data CRM pipeline, task completion, follow-up tracking

Partnership Models

Technology Integration

The natural starting point. WorkingAgents emits OpenTelemetry traces from its MCP dispatcher. DBNL ingests, enriches, and analyzes.

WorkingAgents gains: Behavioral discovery and continuous improvement analytics without building ML infrastructure. Distributional gains: A production MCP reference customer on Elixir/OTP with rich multi-tool, multi-provider agent data.

Consulting Partnership

Distributional’s team comes from SigOpt, Bloomberg, Google, Meta, Stripe, and Uber. They understand enterprise AI deployment. WorkingAgents’ consulting firm deploys agents for medium-size companies. The partnership creates a joint offering:

For Distributional, this is channel distribution through consulting engagements. For WorkingAgents, this is a continuous-improvement service tier that generates recurring revenue.

Co-Marketing: “The Open AI Agent Analytics Stack”

Both companies share a deployment philosophy: open, self-hosted, data stays in your environment. A joint positioning as “the open stack for production AI agents” — orchestration (WorkingAgents) plus analytics (DBNL) — differentiates from cloud-locked alternatives.

Distributional’s $30M in funding from a16z and Two Sigma Ventures gives them marketing reach. A case study showing DBNL discovering behavioral patterns in a production MCP agent platform would be distinctive content for both companies.


Recommended Next Steps

  1. Deploy DBNL sandbox — Distributional offers a free sandbox at docs.dbnl.com. Connect WorkingAgents’ MCP dispatcher traces. See what the platform discovers from even a week of production data.

  2. Instrument the MCP dispatcher — Add OpenTelemetry span emission to MyMCPServer.Manager. Include tool name, parameters, user ID, provider, and session ID as span attributes. This is the minimum data DBNL needs.

  3. Run the flywheel once — Ingest a month of traces. Let DBNL’s unsupervised analysis run. Review the Insights. Pick one discovered pattern and optimize for it. Measure the result. This single cycle demonstrates the value proposition to consulting clients.

  4. Contact Distributional — They’re a Series A company actively expanding. The SigOpt team built their reputation on optimization for enterprise AI. An MCP-native agent orchestration reference on a non-Python stack would be a differentiated story for their portfolio.

  5. Design the consulting package — “Managed AI Agent Operations with Behavioral Analytics” — deploy agents, connect DBNL, deliver monthly intelligence reports, continuously optimize. This is the repeating revenue model.


Conclusion

Distributional solves a problem that most AI teams don’t even know they have: the unknown unknowns in agent behavior. WorkingAgents builds the agents. Distributional discovers what those agents are actually doing in production — the behavioral patterns, correlations, clusters, and drifts that no amount of manual log reading or predefined metrics will surface.

The combination is particularly powerful for consulting. Walk into a client meeting and say: “We deploy AI agents, and we use statistical behavioral analysis to discover patterns in how they operate. Last month we found that your procurement agent was using an inefficient tool chain on 23% of requests. We optimized it. Here’s the before-and-after distributional fingerprint.”

That’s not a pitch. That’s evidence.

DBNL is free, open, and deploys in your infrastructure. The integration is OpenTelemetry — protocol-level, language-agnostic. The team is ex-SigOpt, Google, Meta, Bloomberg. The funding is a16z and Two Sigma. And they’re still early enough that a partnership conversation gets real attention.

The flywheel starts with one trace. Time to emit it.


Sources: