Recommendation: use a hybrid model, but do not price raw “requests” the way LLM vendors do. For an orchestrator like workingagents.ai, the cleanest structure is:

Recommendation: use a hybrid model, but do not price raw “requests” the way LLM vendors do. For an orchestrator like workingagents.ai, the cleanest structure is:

  1. Platform fee
  2. Usage fee based on weighted executions/credits
  3. Optional implementation/consulting at your hourly rate

$300/hour for integration and custom interface work is reasonable if you are doing senior-level API integration, workflow design, security, and custom MCP/tool work. Keep that separate from product pricing so buyers understand they are paying for delivery, not just runtime.

For the product itself, I’d avoid “per request” pricing because one request can vary wildly in cost and value. Market comps are closer to workflow runs / tasks / credits:

My recommended pricing model:

Define credits by cost intensity, not by HTTP request count:

For LLM usage, I would either:

That is better than pretending all orchestrator requests are equivalent.

Consulting recommendation:

Why this works: