AgencyRelay
Capability · AI Agents

White-label AI agents that take real actions

Production agents that call tools, run multi-step workflows, and have a fallback story when the model is wrong — built inside your delivery process and behind your brand.

  • Tool-using, multi-step, production-grade
  • Evals + guardrails before any user sees it
  • White-label safe by default
What an AI Agents engagement looks likeCapability

Agent pod, predictable spine

  • FormatAgent pod inside your SOW
  • CadenceWeekly delivery · async daily standup
  • StackOpenAI · Anthropic · LangChain · custom orchestration
  • Starting at$3,200 / week
Final SOW is scoped against your brief. Multi-track AI pods (e.g. Agents + RAG) and pods that mix capabilities are quoted at the highest applicable rate.
When agencies bring us in

Four moments where the agent brief deserves more than a Friday-night demo

These are the conversations agency owners describe when an agent brief is in front of them and the in-house team has shipped one orchestrator script but not yet a system that survives a billing-day incident.

Signal 0101 / 04

An agent demo got applause and now production wants real guarantees

The video looked great. The buyer is quoting it back at you. Production wants p95 latency, retries, evals, observability, a cost ceiling, and a written fallback when the agent picks the wrong tool. We re-architect the demo into a system that survives a Monday-morning audit.

Signal 0202 / 04

An agent is in a sales-side workflow and confidently does the wrong thing

It books the wrong meeting, drafts the wrong reply, fires the wrong webhook. We add evals against a labelled task set, scoped tool permissions, and a guardrail layer — without throwing away the working autonomy.

Signal 0303 / 04

The team built one agent and the buyer wants three

Each new agent multiplies the prompt, eval, observability, and incident surface. We come in with an orchestration pattern and a per-agent runbook so the team can ship the next two without doubling the on-call burden.

Signal 0404 / 04

An agent platform is being chosen and the choice keeps slipping

OpenAI Agents SDK, LangGraph, Anthropic computer-use, Vercel AI SDK, custom — the trade-offs are real and the team keeps re-litigating. We pick the platform inside the architecture readout, write the trade-offs down, and move to build inside the same week.

What this track is — and isn't

A senior agent build pod under your brand. Not a chat surface, not a single-prompt experiment.

What it covers
  • Tool-using agents wired to your client's existing stack (CRM, helpdesk, internal APIs)
  • Multi-step orchestration with state, retries, and human-in-the-loop checkpoints
  • Eval suite against a labelled task set, run in CI on every prompt change
  • Guardrails — scoped tool permissions, content filters, output-shape validation
  • Observability for run-time behaviour, cost, latency, and failure modes
  • Production deployment with a documented fallback when the agent is uncertain
What it doesn't do
  • Stand-alone chatbots without an action surface — that's often a RAG or Integrations brief
  • Internal automation pipelines without a model in the loop — that's Workflow Automation
  • AI inside an existing CRM or helpdesk surface — that's AI Integrations
  • Direct-to-client pitching — the pod sits inside your team, not in front of the client
  • Recruit, place, or staff-augment a developer onto your payroll
How an agent engagement runs

From brief to first evaluated agent in under three weeks

Agent work runs eval-first. The pod doesn't ship a tool-using agent into a real workflow without a labelled task set behind it.

  1. Step 01Days 1–4

    Brief & feasibility

    Working session with your delivery lead and the buyer-side stakeholder. We pressure-test the brief — what the agent must do, which tools it can call, what "wrong" looks like, and the cost ceiling. NDA and SOW signed under Salt Technologies, Inc.

  2. Step 02Week 1

    Architecture + eval set

    Architecture readout: chosen platform (OpenAI Agents SDK, LangGraph, custom), tool surface, state model, observability, guardrails. In parallel, we build the labelled eval set every prompt and tool change runs against in CI.

  3. Step 03Weeks 2–6

    Build & weekly review

    Iterative agent build with weekly working review. Eval pass rate, run-time cost, and latency tracked from build one. Guardrails and human-in-the-loop checkpoints land before the agent ever touches a real system of record.

  4. Step 04Week 6+

    Production rollout

    Production deployment behind a feature flag, with a percentage-rollout plan and a written runbook for incidents. Post-launch support transitions cleanly into a Support & Maintenance retainer for model drift, regression on new model versions, and v1.x feature work.

How to engage

Two engagement shapes — pick the one that matches your brief

Agent work is long-arc. The two shapes below are the engagements where production-grade agent systems actually land — a new service line under your brand, or a delivery pod inside an active client SOW.

Capability rate
$3,200per week

Starting weekly rate for a single-capability AI pod. Multi-track AI pods (Agents + RAG, Agents + Integrations) and pods that mix capabilities are quoted at the highest applicable rate. Final SOW is scoped against the brief.

Stack & deliverables

Senior AI engineers, your tools, ship-ready output

We work inside the agent tooling your team and your client already use — no parallel platform, no "we'll just rebuild it our way" surprise.

Agent platforms
  • OpenAI Agents SDK
  • LangChain · LangGraph
  • Anthropic SDK + tool use
  • Vercel AI SDK
Tools & state
  • Custom tool wrappers
  • Persistent state (Redis · Postgres)
  • Queues + retries
  • Human-in-the-loop checkpoints
Evals & observability
  • Promptfoo · Braintrust
  • Labelled task sets
  • LangSmith · Helicone · OpenTelemetry
  • Cost dashboards + alerts
Outputs we ship
  • Production agent system
  • Eval suite + CI integration
  • Guardrails + tool permission scopes
  • Observability dashboard
  • Cost monitoring + alerting
  • Incident runbook for your team
Operating principles

Partner-safe inside your top agent accounts

Every AI Agents engagement runs on the same operating spine that protects long-arc retainers and Dedicated Partner Pods — contracted through Salt Technologies, Inc.

Principle

No client-facing footprint

We don't email your client, join their calls, or appear in the proposal — unless you explicitly white-list a named engineer in the SOW.

Principle

Inside your accounts

We work in your GitHub, your model-provider accounts, your hosting, and your shared channel under aliases that fit your team's naming.

Principle

Mutual no-poach

Mutual non-solicitation written into every MSA, with a defined window after the engagement ends. Same clause across every track.

Principle

Salt Technologies, Inc.

MSA, NDA, and engagement SOW are issued by Salt — the Delaware C-Corp behind AgencyRelay.

The same operating spine sits underneath every AgencyRelay capability. Read the no-poach and confidentiality page for the contractual instruments behind these defaults.

AI Agents FAQ

What agency owners ask before sizing an agent build

Direct answers to the questions that come up on almost every AI Agents scoping call.

See full FAQ
  • Q.01

    What's the difference between AI Agents and RAG on this site?

    RAG grounds an answer in your client's content — the model retrieves and cites, but doesn't act. Agents take actions — they call tools, update systems of record, kick off workflows. Most agentic systems we ship combine both (an agent with a RAG tool in its toolbox), but the *primary* capability on the SOW is the one that defines the system. We make this call inside the brief & feasibility step in week one.

  • Q.02

    How do you stop an agent from doing the wrong thing in production?

    Three layers. First, scoped tool permissions — the agent literally can't call tools it isn't authorised for. Second, an eval suite that runs in CI on every prompt or tool change against a labelled task set the buyer signs off on. Third, human-in-the-loop checkpoints on irreversible actions (sending money, deleting data, contacting a customer). All three are part of default scope; we don't ship without them.

  • Q.03

    Which agent platform do you use?

    We don't have a single default. The choice between OpenAI Agents SDK, LangGraph, Anthropic tool use, Vercel AI SDK, and a custom orchestration sits on platform maturity, your team's existing stack, and the latency / cost envelope for the workload. The architecture readout in week one names the platform with the trade-offs written down so your team and your client can refer back.

  • Q.04

    Whose API keys and model-provider accounts do you use?

    By default the client's. We work inside their OpenAI, Anthropic, or other model-provider accounts under named seats. Where you (the agency) hold the account on behalf of the client, we operate inside yours. Either way, the account ownership and cost-attribution boundary is documented in the SOW so there's no ambiguity at month-end.

  • Q.05

    How do you handle agent cost run-away?

    Cost ceiling per agent run, alerting on per-day spend deltas, and a circuit-breaker that pauses the agent surface above a threshold the buyer sets. Cost dashboards land before the production rollout step, not after. We've seen enough cost incidents in our previous lives to design for them up front.

  • Q.06

    What's the smallest engagement you'll take?

    Production agent systems aren't a one-week capability. The most common starting shape is a 4–6 week pod inside an Invisible Delivery Team SOW, sized around the live brief and the integration surface. For shorter agent work (a tightly scoped feature, an eval-suite stand-up, a guardrail audit), we'll quote a tighter window against the same weekly rate.

  • Q.07

    How does the pricing work for a multi-track or multi-capability AI pod?

    The starting weekly rate for a single-capability AI pod is $3,200 per week. Multi-track AI pods (Agents + RAG, Agents + Integrations) and pods that mix capabilities (Agents + UI/UX, Agents + Backend) are quoted at the highest applicable rate. Final SOW is scoped against the brief; the rate is the floor, not a ceiling.

  • Q.08

    What's the right way to support an agent system after launch?

    Most agent systems graduate cleanly into a Support & Maintenance retainer post-launch — model drift monitoring, eval regression on new model versions, guardrail tuning, and v1.x feature work inside a monthly envelope. Either the same pod or a smaller maintenance crew carries it on the same MSA, no second sales cycle.

  • Q.09

    Do we own the work the pod produces?

    Yes. IP ownership and assignment on delivered code, prompts, evals, tool wrappers, and supporting artefacts is written into the MSA — the work belongs to your agency (and onwards to your client per your own client contract) on payment of the relevant invoice. The Salt Technologies templates are counsel-reviewed and shared before signing.

Bring the brief, get the right shape

Tell us the agent work you're sizing — we'll respond with a clean read on platform, pod shape, and starting rate.

Sales-side triage, support deflection, internal ops, or a client-facing agent surface. Either way, the conversation starts with the work — not with a deck.

Operating defaultsMSA / NDA / SOW issued by Salt Technologies, Inc.US-aligned working hoursNo-poach commitmentsWhite-label safe by default