Q.01
What's the difference between RAG and AI Agents on this site?
RAG grounds an answer in your client's content — the model retrieves and cites, but doesn't act. Agents take actions — they call tools, update systems of record, kick off workflows. Most agentic systems we ship combine both (an agent with a RAG tool in its toolbox), but the *primary* capability on the SOW is the one that defines the system. We make this call inside the brief & feasibility step in week one.
Q.02
How do you stop the model from making things up?
Three layers. First, a citation contract — the model is constrained to answer from retrieved chunks and surface the source for each claim. Second, an evaluated "I don't know" path that fires when retrieval is empty or low-confidence, instead of letting the model reach for its training data. Third, an eval suite that runs in CI on every retrieval or prompt change against a labelled question set the buyer signs off on, including hallucination scoring. All three are part of default scope.
Q.03
Which vector store and embedding model do you use?
We don't have a single default. The choice between pgvector (when the corpus is small-to-medium and Postgres is already in the stack) and a managed vector store like Pinecone, Weaviate, or Qdrant (for scale, hybrid filters, or multi-tenant routing) sits on corpus size, latency budget, and your client's existing data layer. Embedding choice — OpenAI text-embedding-3, Voyage, Cohere — is selected in the architecture readout with the trade-offs written down so your team and your client can refer back.
Q.04
How do you handle the corpus — who owns it, who maintains it?
The corpus and the ingestion pipeline belong to your client. We build the ingestion, normalisation, and chunking layer once, with per-source tuning, and ship a re-index runbook so your client's team (or a Support & Maintenance retainer crew) can keep it fresh. Source-of-truth lives in your client's existing systems — Notion, Drive, Confluence, CMS, ticketing — and the vector store is a derived index, not a parallel content system.
Q.05
What happens when documents change or go stale?
Two patterns, picked in the architecture readout. For corpora that change daily — product docs, support content, CMS — we wire incremental re-indexing on source events (webhooks, polling, or change feeds). For slower-moving corpora — policy, legal, archived knowledge — we ship a scheduled re-index runbook with a freshness dashboard so stale chunks are visible before they become wrong answers. Either way, citation rendering shows the source's last-updated timestamp so users see freshness alongside relevance.
Q.06
What's the smallest engagement you'll take?
Production RAG systems aren't a one-week capability. The most common starting shape is a 4–6 week pod inside an Invisible Delivery Team SOW, sized around the corpus surface and the citation contract. For shorter retrieval work (a tightly scoped feature, an eval-suite stand-up, a chunking-strategy audit), we'll quote a tighter window against the same weekly rate.
Q.07
How does the pricing work for a multi-track or multi-capability AI pod?
The starting weekly rate for a single-capability AI pod is $3,200 per week. Multi-track AI pods (RAG + Agents, RAG + Integrations) and pods that mix capabilities (RAG + UI/UX, RAG + Backend) are quoted at the highest applicable rate. Final SOW is scoped against the brief; the rate is the floor, not a ceiling.
Q.08
What's the right way to support a RAG system after launch?
Most RAG systems graduate cleanly into a Support & Maintenance retainer post-launch — corpus drift monitoring, eval regression on new model versions, reranker re-tuning, citation-accuracy review, and v1.x feature work inside a monthly envelope. Either the same pod or a smaller maintenance crew carries it on the same MSA, no second sales cycle.
Q.09
Do we own the work the pod produces?
Yes. IP ownership and assignment on delivered code, prompts, evals, ingestion pipelines, retrieval logic, and supporting artefacts is written into the MSA — the work belongs to your agency (and onwards to your client per your own client contract) on payment of the relevant invoice. The Salt Technologies templates are counsel-reviewed and shared before signing.